Any two eigen functions belonging to unequal eigen values of a self adjoint operator are

In mathematics, a self-adjoint operator on an infinite-dimensional complex vector space V with inner product ⟨ ⋅ , ⋅ ⟩ {\displaystyle \langle \cdot ,\cdot \rangle }

Any two eigen functions belonging to unequal eigen values of a self adjoint operator are
(equivalently, a Hermitian operator in the finite-dimensional case) is a linear map A (from V to itself) that is its own adjoint. If V is finite-dimensional with a given orthonormal basis, this is equivalent to the condition that the matrix of A is a Hermitian matrix, i.e., equal to its conjugate transpose A∗. By the finite-dimensional spectral theorem, V has an orthonormal basis such that the matrix of A relative to this basis is a diagonal matrix with entries in the real numbers. In this article, we consider generalizations of this concept to operators on Hilbert spaces of arbitrary dimension.

Self-adjoint operators are used in functional analysis and quantum mechanics. In quantum mechanics their importance lies in the Dirac–von Neumann formulation of quantum mechanics, in which physical observables such as position, momentum, angular momentum and spin are represented by self-adjoint operators on a Hilbert space. Of particular significance is the Hamiltonian operator H ^ {\displaystyle {\hat {H}}}

Any two eigen functions belonging to unequal eigen values of a self adjoint operator are
defined by

H ^ ψ = − ℏ 2 2 m ∇ 2 ψ + V ψ , {\displaystyle {\hat {H}}\psi =-{\frac {\hbar ^{2}}{2m}}\nabla ^{2}\psi +V\psi ,}
Any two eigen functions belonging to unequal eigen values of a self adjoint operator are

which as an observable corresponds to the total energy of a particle of mass m in a real potential field V. Differential operators are an important class of unbounded operators.

The structure of self-adjoint operators on infinite-dimensional Hilbert spaces essentially resembles the finite-dimensional case. That is to say, operators are self-adjoint if and only if they are unitarily equivalent to real-valued multiplication operators. With suitable modifications, this result can be extended to possibly unbounded operators on infinite-dimensional spaces. Since an everywhere-defined self-adjoint operator is necessarily bounded, one needs be more attentive to the domain issue in the unbounded case. This is explained below in more detail.

Let A {\displaystyle A}   be an unbounded (i.e. not necessarily bounded) operator with a dense domain Dom ⁡ A ⊆ H . {\displaystyle \operatorname {Dom} A\subseteq H.}   This condition holds automatically when H {\displaystyle H}   is finite-dimensional since Dom ⁡ A = H {\displaystyle \operatorname {Dom} A=H}   for every linear operator on a finite-dimensional space.

Let the inner product ⟨ ⋅ , ⋅ ⟩ {\displaystyle \langle \cdot ,\cdot \rangle }   be conjugate-linear on the second argument. This applies to complex Hilbert spaces only. By definition, the adjoint operator A ∗ {\displaystyle A^{*}}   acts on the subspace Dom ⁡ A ∗ ⊆ H {\displaystyle \operatorname {Dom} A^{*}\subseteq H}   consisting of the elements y {\displaystyle y}   for which there is a z ∈ H {\displaystyle z\in H}   such that ⟨ A x , y ⟩ = ⟨ x , z ⟩ , {\displaystyle \langle Ax,y\rangle =\langle x,z\rangle ,}   for every x ∈ Dom ⁡ A . {\displaystyle x\in \operatorname {Dom} A.}   Setting A ∗ y = z {\displaystyle A^{*}y=z}   defines the linear operator A ∗ . {\displaystyle A^{*}.}  

The graph of an (arbitrary) operator A {\displaystyle A}   is the set G ( A ) = { ( x , A x ) ∣ x ∈ Dom ⁡ A } . {\displaystyle G(A)=\{(x,Ax)\mid x\in \operatorname {Dom} A\}.}   An operator B {\displaystyle B}   is said to extend A {\displaystyle A}   if G ( A ) ⊆ G ( B ) . {\displaystyle G(A)\subseteq G(B).}   This is written as A ⊆ B . {\displaystyle A\subseteq B.}  

The densely defined operator A {\displaystyle A}   is called symmetric if

⟨ A x , y ⟩ = ⟨ x , A y ⟩ , {\displaystyle \langle Ax,y\rangle =\langle x,Ay\rangle ,}  

for all x , y ∈ Dom ⁡ A . {\displaystyle x,y\in \operatorname {Dom} A.}   As shown below, A {\displaystyle A}   is symmetric if and only if G ( A ) ⊆ G ( A ∗ ) . {\displaystyle G(A)\subseteq G(A^{*}).}  

The unbounded densely defined operator A {\displaystyle A}   is called self-adjoint if G ( A ) = G ( A ∗ ) . {\displaystyle G(A)=G(A^{*}).}   Explicitly, Dom ⁡ A = Dom ⁡ A ∗ {\displaystyle \operatorname {Dom} A=\operatorname {Dom} A^{*}}   and A = A ∗ . {\displaystyle A=A^{*}.}   Every self-adjoint operator is symmetric. Conversely, a symmetric operator A {\displaystyle A}   for which Dom ⁡ A = Dom ⁡ A ∗ {\displaystyle \operatorname {Dom} A=\operatorname {Dom} A^{*}}   is self-adjoint. In physics, the term Hermitian refers to symmetric as well as self-adjoint operators alike. The subtle difference between the two is generally overlooked.

A subset ρ ( A ) ⊆ C {\displaystyle \rho (A)\subseteq \mathbb {C} }   is called the resolvent set (or regular set) if for every λ ∈ ρ ( A ) , {\displaystyle \lambda \in \rho (A),}   the (not-necessarily-bounded) operator A − λ I {\displaystyle A-\lambda I}   has a bounded everywhere-defined inverse. The complement σ ( A ) = C ∖ ρ ( A ) {\displaystyle \sigma (A)=\mathbb {C} \setminus \rho (A)}   is called spectrum. In finite dimensions, σ ( A ) {\displaystyle \sigma (A)}   consists exclusively of eigenvalues.

A bounded operator A is self-adjoint if

⟨ A x , y ⟩ = ⟨ x , A y ⟩ {\displaystyle \langle Ax,y\rangle =\langle x,Ay\rangle }  

for all x {\displaystyle x}   and y {\displaystyle y}   in H. If A is symmetric and D o m ( A ) = H {\displaystyle \mathrm {Dom} (A)=H}  , then, by Hellinger–Toeplitz theorem, A is necessarily bounded.[1]

Every bounded linear operator T : H → H on a Hilbert space H can be written in the form T = A + i B {\displaystyle T=A+iB}   where A : H → H and B : H → H are bounded self-adjoint operators.[2]

Properties of bounded self-adjoint operators

Let H be a Hilbert space and let A : H → H {\displaystyle A:H\to H}   be a bounded self-adjoint linear operator defined on D ⁡ ( A ) = H {\displaystyle \operatorname {D} \left(A\right)=H}  .

  • ⟨ h , A h ⟩ {\displaystyle \left\langle h,Ah\right\rangle }   is real for all h ∈ H {\displaystyle h\in H}  .[3]
  • ‖ A ‖ = sup { | ⟨ h , A h ⟩ | : ‖ h ‖ = 1 } {\displaystyle \left\|A\right\|=\sup \left\{|\langle h,Ah\rangle |:\|h\|=1\right\}}  [3] if dim ⁡ H ≠ 0. {\displaystyle \operatorname {dim} H\neq 0.}  
  • If the image of A, denoted by Im ⁡ A {\displaystyle \operatorname {Im} A}  , is dense in H then A : H → Im ⁡ A {\displaystyle A:H\to \operatorname {Im} A}   is invertible.
  • The eigenvalues of A are real and eigenvectors belonging to different eigenvalues are orthogonal.[3]
  • If λ {\displaystyle \lambda }   is an eigenvalue of A then | λ | ≤ ‖ A ‖ {\displaystyle |\lambda |\leq \|A\|}  ; in particular, | λ | ≤ sup { | ⟨ h , A h ⟩ | : ‖ h ‖ ≤ 1 } {\displaystyle |\lambda |\leq \sup \left\{|\langle h,Ah\rangle |:\|h\|\leq 1\right\}}  .[3]
    • In general, there may not exist any eigenvalue λ {\displaystyle \lambda }   such that | λ | = sup { | ⟨ h , A h ⟩ | : ‖ h ‖ ≤ 1 } {\displaystyle |\lambda |=\sup \left\{|\langle h,Ah\rangle |:\|h\|\leq 1\right\}}  , but if in addition A is compact then there necessarily exists an eigenvalue λ {\displaystyle \lambda }  , equal to either ‖ A ‖ {\displaystyle \|A\|}   or − ‖ A ‖ {\displaystyle -\|A\|}  ,[4] such that | λ | = sup { | ⟨ h , A h ⟩ | : ‖ h ‖ ≤ 1 } {\displaystyle |\lambda |=\sup \left\{|\langle h,Ah\rangle |:\|h\|\leq 1\right\}}  ,[3]
  • If a sequence of bounded self-adjoint linear operators is convergent then the limit is self-adjoint.[2]
  • There exists a number λ {\displaystyle \lambda }  , equal to either ‖ A ‖ {\displaystyle \|A\|}   or − ‖ A ‖ {\displaystyle -\|A\|}  , and a sequence ( x i ) i = 1 ∞ ⊆ H {\displaystyle \left(x_{i}\right)_{i=1}^{\infty }\subseteq H}   such that lim i → ∞ A x i − λ x i = 0 {\displaystyle \lim _{i\to \infty }Ax_{i}-\lambda x_{i}=0}   and ‖ x i ‖ = 1 {\displaystyle \|x_{i}\|=1}   for all i.[4]

NOTE: symmetric operators are defined above.

A is symmetric ⇔ A⊆A*

An unbounded, densely defined operator A {\displaystyle A}   is symmetric if and only if A ⊆ A ∗ . {\displaystyle A\subseteq A^{*}.}   Indeed, the if-part follows directly from the definition of the adjoint operator. For the only-if-part, assuming that A {\displaystyle A}   is symmetric, the inclusion Dom ⁡ ( A ) ⊆ Dom ⁡ ( A ∗ ) {\displaystyle \operatorname {Dom} (A)\subseteq \operatorname {Dom} (A^{*})}   follows from the Cauchy–Bunyakovsky–Schwarz inequality: for every x , y ∈ Dom ⁡ ( A ) , {\displaystyle x,y\in \operatorname {Dom} (A),}  

| ⟨ A x , y ⟩ | = | ⟨ x , A y ⟩ | ≤ ‖ x ‖ ⋅ ‖ A y ‖ . {\displaystyle |\langle Ax,y\rangle |=|\langle x,Ay\rangle |\leq \|x\|\cdot \|Ay\|.}  

The equality A = A ∗ | Dom ⁡ ( A ) {\displaystyle A=A^{*}|_{\operatorname {Dom} (A)}}   holds due to the equality

⟨ x , A ∗ y ⟩ = ⟨ A x , y ⟩ = ⟨ x , A y ⟩ , {\displaystyle \langle x,A^{*}y\rangle =\langle Ax,y\rangle =\langle x,Ay\rangle ,}  

for every x , y ∈ Dom ⁡ A ⊆ Dom ⁡ A ∗ , {\displaystyle x,y\in \operatorname {Dom} A\subseteq \operatorname {Dom} A^{*},}   the density of Dom ⁡ A , {\displaystyle \operatorname {Dom} A,}   and non-degeneracy of the inner product.

The Hellinger–Toeplitz theorem says that an everywhere-defined symmetric operator is bounded and self-adjoint.

A is symmetric ⇔ ∀x ⟨Ax, x⟩ ∈ R

The only-if part follows directly from the definition (see above). To prove the if-part, assume without loss of generality that the inner product ⟨ ⋅ , ⋅ ⟩ {\displaystyle \langle \cdot ,\cdot \rangle }   is anti-linear on the first argument and linear on the second. (In the reverse scenario, we work with ⟨ x , y ⟩ op = def   ⟨ y , x ⟩ {\displaystyle \langle x,y\rangle _{\text{op}}{\stackrel {\text{def}}{=}}\ \langle y,x\rangle }   instead). The symmetry of A {\displaystyle A}   follows from the polarization identity

⟨ A x , y ⟩ = ⟨ A ( x + y ) , x + y ⟩ − ⟨ A ( x − y ) , x − y ⟩ − i ⟨ A ( x + i y ) , x + i y ⟩ + i ⟨ A ( x − i y ) , x − i y ⟩ {\displaystyle {\begin{aligned}\langle Ax,y\rangle ={}&\langle A(x+y),x+y\rangle -\langle A(x-y),x-y\rangle \\[1mm]&{}-i\langle A(x+iy),x+iy\rangle +i\langle A(x-iy),x-iy\rangle \end{aligned}}}  

which holds for every x , y ∈ Dom ⁡ A . {\displaystyle x,y\in \operatorname {Dom} A.}  

||(A−λ)x|| ≥ d(λ)⋅||x||

This property is used in the proof that the spectrum of a self-adjoint operator is real.

Define S = { x ∈ Dom ⁡ A ∣ ‖ x ‖ = 1 } , {\displaystyle S=\{x\in \operatorname {Dom} A\mid \Vert x\Vert =1\},}   m = inf x ∈ S ⟨ A x , x ⟩ , {\displaystyle \textstyle m=\inf _{x\in S}\langle Ax,x\rangle ,}   and M = sup x ∈ S ⟨ A x , x ⟩ . {\displaystyle \textstyle M=\sup _{x\in S}\langle Ax,x\rangle .}   The values m , M ∈ R ∪ { ± ∞ } {\displaystyle m,M\in \mathbb {R} \cup \{\pm \infty \}}   are properly defined since S ≠ ∅ , {\displaystyle S\neq \emptyset ,}   and ⟨ A x , x ⟩ ∈ R , {\displaystyle \langle Ax,x\rangle \in \mathbb {R} ,}   due to symmetry. Then, for every λ ∈ C {\displaystyle \lambda \in \mathbb {C} }   and every x ∈ Dom ⁡ A , {\displaystyle x\in \operatorname {Dom} A,}  

‖ A − λ x ‖ ≥ d ( λ ) ⋅ ‖ x ‖ , {\displaystyle \Vert A-\lambda x\Vert \geq d(\lambda )\cdot \Vert x\Vert ,}  

where d ( λ ) = inf r ∈ [ m , M ] | r − λ | . {\displaystyle \textstyle d(\lambda )=\inf _{r\in [m,M]}|r-\lambda |.}  

Indeed, let x ∈ Dom ⁡ A ∖ { 0 } . {\displaystyle x\in \operatorname {Dom} A\setminus \{0\}.}   By Cauchy-Schwarz inequality,

‖ A − λ x ‖ ≥ | ⟨ A − λ x , x ⟩ | ‖ x ‖ = | ⟨ A x ‖ x ‖ , x ‖ x ‖ ⟩ − λ | ⋅ ‖ x ‖ ≥ d ( λ ) ⋅ ‖ x ‖ . {\displaystyle \Vert A-\lambda x\Vert \geq {\frac {|\langle A-\lambda x,x\rangle |}{\Vert x\Vert }}=\left|\left\langle A{\frac {x}{\Vert x\Vert }},{\frac {x}{\Vert x\Vert }}\right\rangle -\lambda \right|\cdot \Vert x\Vert \geq d(\lambda )\cdot \Vert x\Vert .}  

If λ ∉ [ m , M ] , {\displaystyle \lambda \notin [m,M],}   then d ( λ ) > 0 , {\displaystyle d(\lambda )>0,}   and A − λ I {\displaystyle A-\lambda I}   is called bounded below.

A simple example

As noted above, the spectral theorem applies only to self-adjoint operators, and not in general to symmetric operators. Nevertheless, we can at this point give a simple example of a symmetric operator that has an orthonormal basis of eigenvectors. (This operator is actually "essentially self-adjoint.") The operator A below can be seen to have a compact inverse, meaning that the corresponding differential equation Af = g is solved by some integral, therefore compact, operator G. The compact symmetric operator G then has a countable family of eigenvectors which are complete in L2. The same can then be said for A.

Consider the complex Hilbert space L2[0,1] and the differential operator

A = − d 2 d x 2 {\displaystyle A=-{\frac {d^{2}}{dx^{2}}}}  

with D o m ( A ) {\displaystyle \mathrm {Dom} (A)}   consisting of all complex-valued infinitely differentiable functions f on [0, 1] satisfying the boundary conditions

f ( 0 ) = f ( 1 ) = 0. {\displaystyle f(0)=f(1)=0.}  

Then integration by parts of the inner product shows that A is symmetric. The reader is invited to perform integration by parts twice and verify that the given boundary conditions for Dom ⁡ ( A ) {\displaystyle \operatorname {Dom} (A)}   ensure that the boundary terms in the integration by parts vanish.

The eigenfunctions of A are the sinusoids

f n ( x ) = sin ⁡ ( n π x ) n = 1 , 2 , … {\displaystyle f_{n}(x)=\sin(n\pi x)\qquad n=1,2,\ldots }  

with the real eigenvalues n2π2; the well-known orthogonality of the sine functions follows as a consequence of the property of being symmetric.

We consider generalizations of this operator below.

Let A {\displaystyle A}   be an unbounded symmetric operator. A {\displaystyle A}   is self-adjoint if and only if σ ( A ) ⊆ R . {\displaystyle \sigma (A)\subseteq \mathbb {R} .}  

Proof: self-adjoint operator has real spectrum

Let A {\displaystyle A}   be self-adjoint. Self-adjoint operators are symmetric. The initial steps of this proof are carried out based on the symmetry alone. Self-adjointness of A {\displaystyle A}   is not used directly until step 1b(i). Let λ ∈ C . {\displaystyle \lambda \in \mathbb {C} .}   Denote R λ = A − λ I . {\displaystyle R_{\lambda }=A-\lambda I.}   Using the notations from the section on symmetric operators (see above), it suffices to prove that σ ( A ) ⊆ [ m , M ] . {\displaystyle \sigma (A)\subseteq [m,M].}  

  1. Let λ ∈ C ∖ [ m , M ] . {\displaystyle \lambda \in \mathbb {C} \setminus [m,M].}   The goal is to prove the existence and boundedness of the inverted resolvent operator R λ − 1 , {\displaystyle R_{\lambda }^{-1},}   and show that Dom ⁡ R λ − 1 = H . {\displaystyle \operatorname {Dom} R_{\lambda }^{-1}=H.}   We begin by showing that ker ⁡ R λ = { 0 } {\displaystyle \ker R_{\lambda }=\{0\}}   and Im ⁡ R λ = H . {\displaystyle \operatorname {Im} R_{\lambda }=H.}  

    1. As shown above, R λ {\displaystyle R_{\lambda }}   is bounded below, i.e. ‖ R λ x ‖ ≥ d ( λ ) ⋅ ‖ x ‖ , {\displaystyle \Vert R_{\lambda }x\Vert \geq d(\lambda )\cdot \Vert x\Vert ,}   with d ( λ ) > 0. {\displaystyle d(\lambda )>0.}   The triviality of ker ⁡ R λ {\displaystyle \ker R_{\lambda }}   follows.
    2. It remains to show that Im ⁡ R λ = H . {\displaystyle \operatorname {Im} R_{\lambda }=H.}   Indeed,

      1. Im ⁡ R λ {\displaystyle \operatorname {Im} R_{\lambda }}   is closed. To prove this, pick a sequence y n = R λ x n ∈ Im ⁡ R λ {\displaystyle y_{n}=R_{\lambda }x_{n}\in \operatorname {Im} R_{\lambda }}   converging to some y ∈ H . {\displaystyle y\in H.}   Since

        ‖ x n − x m ‖ ≤ 1 d ( λ ) ‖ y n − y m ‖ , {\displaystyle \|x_{n}-x_{m}\|\leq {\frac {1}{d(\lambda )}}\|y_{n}-y_{m}\|,}

         

        x n {\displaystyle x_{n}}   is fundamental. Hence, it converges to some x ∈ H . {\displaystyle x\in H.}   Furthermore, y n + λ x n = A x n {\displaystyle y_{n}+\lambda x_{n}=Ax_{n}}   and y n + λ x n → y + λ x . {\displaystyle y_{n}+\lambda x_{n}\to y+\lambda x.}   One should emphasize that the arguments made thus far hold for any symmetric but not necessarily self-adjoint operator. It now follows from self-adjointness that A {\displaystyle A}   is closed, so x ∈ Dom ⁡ A = Dom ⁡ R λ , {\displaystyle x\in \operatorname {Dom} A=\operatorname {Dom} R_{\lambda },}   A x = y + λ x ∈ Im ⁡ A , {\displaystyle Ax=y+\lambda x\in \operatorname {Im} A,}   and consequently y = R λ x ∈ Im ⁡ R λ . {\displaystyle y=R_{\lambda }x\in \operatorname {Im} R_{\lambda }.}   Finally,
      2. Im ⁡ R λ {\displaystyle \operatorname {Im} R_{\lambda }}   is dense in H . {\displaystyle H.}   Indeed, the article about Adjoint operator points out that Im ⁡ R λ ⊥ = ker ⁡ R λ ∗ . {\displaystyle \operatorname {Im} R_{\lambda }^{\perp }=\ker R_{\lambda }^{*}.}   From self-adjointness of A {\displaystyle A}   (i.e. A ∗ = A ) {\displaystyle A^{*}=A)}  , R λ ∗ = R λ ¯ . {\displaystyle R_{\lambda }^{*}=R_{\bar {\lambda }}.}   Since λ ∈ C ∖ [ m , M ] , {\displaystyle \lambda \in \mathbb {C} \setminus [m,M],}   the inclusion λ ¯ ∈ C ∖ [ m , M ] {\displaystyle {\bar {\lambda }}\in \mathbb {C} \setminus [m,M]}   implies that d ( λ ¯ ) > 0 , {\displaystyle d({\bar {\lambda }})>0,}   and consequently, ker ⁡ R λ ¯ = { 0 } . {\displaystyle \ker R_{\bar {\lambda }}=\{0\}.}  

  2. The operator R λ : Dom ⁡ A → H {\displaystyle R_{\lambda }\colon \operatorname {Dom} A\to H}   has now been proven to be bijective, so the set-theoretic inverse R λ − 1 {\displaystyle R_{\lambda }^{-1}}   exists and is everywhere defined. The graph of R λ − 1 {\displaystyle R_{\lambda }^{-1}}   is the set { ( R λ x , x ) ∣ x ∈ Dom ⁡ A } . {\displaystyle \{(R_{\lambda }x,x)\mid x\in \operatorname {Dom} A\}.}   Since R λ {\displaystyle R_{\lambda }}   is closed (because A {\displaystyle A}   is), so is R λ − 1 . {\displaystyle R_{\lambda }^{-1}.}   By closed graph theorem, R λ − 1 {\displaystyle R_{\lambda }^{-1}}   is bounded, so λ ∉ σ ( A ) . {\displaystyle \lambda \notin \sigma (A).}  

Proof: Symmetric operator with real spectrum is self-adjoint

  1. By assumption, A {\displaystyle A}   is symmetric; therefore A ⊆ A ∗ . {\displaystyle A\subseteq A^{*}.}   For every λ ∈ C , {\displaystyle \lambda \in \mathbb {C} ,}   A − λ I ⊆ A ∗ − λ I . {\displaystyle A-\lambda I\subseteq A^{*}-\lambda I.}   Let σ ( A ) ⊆ [ m , M ] . {\displaystyle \sigma (A)\subseteq [m,M].}   (These constants are defined in the section on symmetic operators above). If λ ∉ [ m , M ] , {\displaystyle \lambda \notin [m,M],}   then λ ¯ ∉ [ m , M ] . {\displaystyle {\bar {\lambda }}\notin [m,M].}   Since λ {\displaystyle \lambda }   and λ ¯ {\displaystyle {\bar {\lambda }}}   are not in the spectrum, the operators A − λ I , A − λ ¯ I : Dom ⁡ A → H {\displaystyle A-\lambda I,A-{\bar {\lambda }}I:\operatorname {Dom} A\to H}   are bijective. Moreover,
  2. A − λ I = A ∗ − λ I . {\displaystyle A-\lambda I=A^{*}-\lambda I.}   Indeed, H = Im ⁡ ( A − λ I ) ⊆ Im ⁡ ( A ∗ − λ I ) . {\displaystyle H=\operatorname {Im} (A-\lambda I)\subseteq \operatorname {Im} (A^{*}-\lambda I).}   If one had Dom ⁡ ( A − λ I ) ⊊ Dom ⁡ ( A ∗ − λ I ) , {\displaystyle \operatorname {Dom} (A-\lambda I)\subsetneq \operatorname {Dom} (A^{*}-\lambda I),}   then A ∗ − λ I {\displaystyle A^{*}-\lambda I}   would not be injective, i.e. one would have ker ⁡ ( A ∗ − λ I ) ≠ { 0 } . {\displaystyle \ker(A^{*}-\lambda I)\neq \{0\}.}   As discussed in the article about Adjoint operator, Im ⁡ ( A − λ ¯ I ) ⊥ = ker ⁡ ( A ∗ − λ I ) , {\displaystyle \operatorname {Im} (A-{\bar {\lambda }}I)^{\perp }=\ker(A^{*}-\lambda I),}   and, hence, Im ⁡ ( A − λ ¯ I ) ≠ H . {\displaystyle \operatorname {Im} (A-{\bar {\lambda }}I)\neq H.}   This contradicts the bijectiveness.
  3. The equality A − λ I = A ∗ − λ I {\displaystyle A-\lambda I=A^{*}-\lambda I}   shows that A = A ∗ , {\displaystyle A=A^{*},}   i.e. A {\displaystyle A}   is self-adjoint. Indeed, it suffices to prove that A ∗ ⊆ A . {\displaystyle A^{*}\subseteq A.}   For every x ∈ Dom ⁡ A ∗ {\displaystyle x\in \operatorname {Dom} A^{*}}   and y = A ∗ x , {\displaystyle y=A^{*}x,}  

    A ∗ x = y ⇔ ( A ∗ − λ I ) x = y − λ x ⇔ ( A − λ I ) x = y − λ x ⇔ A x = y . {\displaystyle A^{*}x=y\Leftrightarrow (A^{*}-\lambda I)x=y-\lambda x\Leftrightarrow (A-\lambda I)x=y-\lambda x\Leftrightarrow Ax=y.}

     

A symmetric operator A is always closable; that is, the closure of the graph of A is the graph of an operator. A symmetric operator A is said to be essentially self-adjoint if the closure of A is self-adjoint. Equivalently, A is essentially self-adjoint if it has a unique self-adjoint extension. In practical terms, having an essentially self-adjoint operator is almost as good as having a self-adjoint operator, since we merely need to take the closure to obtain self-adjoint operator.

Consider the complex Hilbert space L2(R), and the operator which multiplies a given function by x:

A f ( x ) = x f ( x ) {\displaystyle Af(x)=xf(x)}  

The domain of A is the space of all L2 functions f ( x ) {\displaystyle f(x)}   for which x f ( x ) {\displaystyle xf(x)}   is also square-integrable. Then A is self-adjoint.[5] On the other hand, A does not have any eigenfunctions. (More precisely, A does not have any normalizable eigenvectors, that is, eigenvectors that are actually in the Hilbert space on which A is defined.)

As we will see later, self-adjoint operators have very important spectral properties; they are in fact multiplication operators on general measure spaces.

As has been discussed above, although the distinction between a symmetric operator and a self-adjoint (or essentially self-adjoint) operator is a subtle one, it is important since self-adjointness is the hypothesis in the spectral theorem. Here we discuss some concrete examples of the distinction; see the section below on extensions of symmetric operators for the general theory.

A note regarding domains

Every self-adjoint operator is symmetric. Conversely, every symmetric operator for which Dom ⁡ ( A ∗ ) ⊆ Dom ⁡ ( A ) {\displaystyle \operatorname {Dom} (A^{*})\subseteq \operatorname {Dom} (A)}   is self-adjoint. Symmetric operators for which Dom ⁡ ( A ∗ ) {\displaystyle \operatorname {Dom} (A^{*})}   is strictly greater than Dom ⁡ ( A ) {\displaystyle \operatorname {Dom} (A)}   cannot be self-adjoint.

Boundary conditions

In the case where the Hilbert space is a space of functions on a bounded domain, these distinctions have to do with a familiar issue in quantum physics: One cannot define an operator—such as the momentum or Hamiltonian operator—on a bounded domain without specifying boundary conditions. In mathematical terms, choosing the boundary conditions amounts to choosing an appropriate domain for the operator. Consider, for example, the Hilbert space L 2 ( [ 0 , 1 ] ) {\displaystyle L^{2}([0,1])}   (the space of square-integrable functions on the interval [0,1]). Let us define a "momentum" operator A on this space by the usual formula, setting Planck's constant equal to 1:

A f = − i d f d x . {\displaystyle Af=-i{\frac {df}{dx}}.}  

We must now specify a domain for A, which amounts to choosing boundary conditions. If we choose

Dom ⁡ ( A ) = { smooth functions } , {\displaystyle \operatorname {Dom} (A)=\left\{{\text{smooth functions}}\right\},}  

then A is not symmetric (because the boundary terms in the integration by parts do not vanish).

If we choose

Dom ⁡ ( A ) = { smooth functions f ∣ f ( 0 ) = f ( 1 ) = 0 } , {\displaystyle \operatorname {Dom} (A)=\left\{{\text{smooth functions}}\,f\mid f(0)=f(1)=0\right\},}  

then using integration by parts, one can easily verify that A is symmetric. This operator is not essentially self-adjoint,[6] however, basically because we have specified too many boundary conditions on the domain of A, which makes the domain of the adjoint too big. (This example is discussed also in the "Examples" section below.)

Specifically, with the above choice of domain for A, the domain of the closure A c l {\displaystyle A^{\mathrm {cl} }}   of A is

Dom ⁡ ( A c l ) = { functions  f  with two derivatives in  L 2 ∣ f ( 0 ) = f ( 1 ) = 0 } , {\displaystyle \operatorname {Dom} \left(A^{\mathrm {cl} }\right)=\left\{{\text{functions }}f{\text{ with two derivatives in }}L^{2}\mid f(0)=f(1)=0\right\},}  

whereas the domain of the adjoint A ∗ {\displaystyle A^{*}}   of A is

Dom ⁡ ( A ∗ ) = { functions  f  with two derivatives in  L 2 } . {\displaystyle \operatorname {Dom} \left(A^{*}\right)=\left\{{\text{functions }}f{\text{ with two derivatives in }}L^{2}\right\}.}  

That is to say, the domain of the closure has the same boundary conditions as the domain of A itself, just a less stringent smoothness assumption. Meanwhile, since there are "too many" boundary conditions on A, there are "too few" (actually, none at all in this case) for A ∗ {\displaystyle A^{*}}  . If we compute ⟨ g , A f ⟩ {\displaystyle \langle g,Af\rangle }   for f ∈ Dom ⁡ ( A ) {\displaystyle f\in \operatorname {Dom} (A)}   using integration by parts, then since f {\displaystyle f}   vanishes at both ends of the interval, no boundary conditions on g {\displaystyle g}   are needed to cancel out the boundary terms in the integration by parts. Thus, any sufficiently smooth function g {\displaystyle g}   is in the domain of A ∗ {\displaystyle A^{*}}  , with A ∗ g = − i d g / d x {\displaystyle A^{*}g=-i\,dg/dx}  .[7]

Since the domain of the closure and the domain of the adjoint do not agree, A is not essentially self-adjoint. After all, a general result says that the domain of the adjoint of A c l {\displaystyle A^{\mathrm {cl} }}   is the same as the domain of the adjoint of A. Thus, in this case, the domain of the adjoint of A c l {\displaystyle A^{\mathrm {cl} }}   is bigger than the domain of A c l {\displaystyle A^{\mathrm {cl} }}   itself, showing that A c l {\displaystyle A^{\mathrm {cl} }}   is not self-adjoint, which by definition means that A is not essentially self-adjoint.

The problem with the preceding example is that we imposed too many boundary conditions on the domain of A. A better choice of domain would be to use periodic boundary conditions:

Dom ⁡ ( A ) = { smooth functions f ∣ f ( 0 ) = f ( 1 ) } . {\displaystyle \operatorname {Dom} (A)=\{{\text{smooth functions}}\,f\mid f(0)=f(1)\}.}  

With this domain, A is essentially self-adjoint.[8]

In this case, we can understand the implications of the domain issues for the spectral theorem. If we use the first choice of domain (with no boundary conditions), all functions f β ( x ) = e β x {\displaystyle f_{\beta }(x)=e^{\beta x}}   for β ∈ C {\displaystyle \beta \in \mathbb {C} }   are eigenvectors, with eigenvalues − i β {\displaystyle -i\beta }  , and so the spectrum is the whole complex plane. If we use the second choice of domain (with Dirichlet boundary conditions), A has no eigenvectors at all. If we use the third choice of domain (with periodic boundary conditions), we can find an orthonormal basis of eigenvectors for A, the functions f n ( x ) := e 2 π i n x {\displaystyle f_{n}(x):=e^{2\pi inx}}  . Thus, in this case finding a domain such that A is self-adjoint is a compromise: the domain has to be small enough so that A is symmetric, but large enough so that D ( A ∗ ) = D ( A ) {\displaystyle D(A^{*})=D(A)}  .

Schrödinger operators with singular potentials

A more subtle example of the distinction between symmetric and (essentially) self-adjoint operators comes from Schrödinger operators in quantum mechanics. If the potential energy is singular—particularly if the potential is unbounded below—the associated Schrödinger operator may fail to be essentially self-adjoint. In one dimension, for example, the operator

H ^ := P 2 2 m − X 4 {\displaystyle {\hat {H}}:={\frac {P^{2}}{2m}}-X^{4}}  

is not essentially self-adjoint on the space of smooth, rapidly decaying functions.[9] In this case, the failure of essential self-adjointness reflects a pathology in the underlying classical system: A classical particle with a − x 4 {\displaystyle -x^{4}}   potential escapes to infinity in finite time. This operator does not have a unique self-adjoint, but it does admit self-adjoint extensions obtained by specifying "boundary conditions at infinity". (Since H ^ {\displaystyle {\hat {H}}}   is a real operator, it commutes with complex conjugation. Thus, the deficiency indices are automatically equal, which is the condition for having a self-adjoint extension. See the discussion of extensions of symmetric operators below.)

In this case, if we initially define H ^ {\displaystyle {\hat {H}}}   on the space of smooth, rapidly decaying functions, the adjoint will be "the same" operator (i.e., given by the same formula) but on the largest possible domain, namely

Dom ⁡ ( H ^ ∗ ) = { twice differentiable functions  f ∈ L 2 ( R ) | ( − ℏ 2 2 m d 2 f d x 2 − x 4 f ( x ) ) ∈ L 2 ( R ) } . {\displaystyle \operatorname {Dom} \left({\hat {H}}^{*}\right)=\left\{{\text{twice differentiable functions }}f\in L^{2}(\mathbb {R} )\left|\left(-{\frac {\hbar ^{2}}{2m}}{\frac {d^{2}f}{dx^{2}}}-x^{4}f(x)\right)\in L^{2}(\mathbb {R} )\right.\right\}.}  

It is then possible to show that H ^ ∗ {\displaystyle {\hat {H}}^{*}}   is not a symmetric operator, which certainly implies that H ^ {\displaystyle {\hat {H}}}   is not essentially self-adjoint. Indeed, H ^ ∗ {\displaystyle {\hat {H}}^{*}}   has eigenvectors with pure imaginary eigenvalues,[10][11] which is impossible for a symmetric operator. This strange occurrence is possible because of a cancellation between the two terms in H ^ ∗ {\displaystyle {\hat {H}}^{*}}  : There are functions f {\displaystyle f}   in the domain of H ^ ∗ {\displaystyle {\hat {H}}^{*}}   for which neither d 2 f / d x 2 {\displaystyle d^{2}f/dx^{2}}   nor x 4 f ( x ) {\displaystyle x^{4}f(x)}   is separately in L 2 ( R ) {\displaystyle L^{2}(\mathbb {R} )}  , but the combination of them occurring in H ^ ∗ {\displaystyle {\hat {H}}^{*}}   is in L 2 ( R ) {\displaystyle L^{2}(\mathbb {R} )}  . This allows for H ^ ∗ {\displaystyle {\hat {H}}^{*}}   to be nonsymmetric, even though both d 2 / d x 2 {\displaystyle d^{2}/dx^{2}}   and X 4 {\displaystyle X^{4}}   are symmetric operators. This sort of cancellation does not occur if we replace the repelling potential − x 4 {\displaystyle -x^{4}}   with the confining potential x 4 {\displaystyle x^{4}}  .

Conditions for Schrödinger operators to be self-adjoint or essentially self-adjoint can be found in various textbooks, such as those by Berezin and Shubin, Hall, and Reed and Simon listed in the references.

In the physics literature, the spectral theorem is often stated by saying that a self-adjoint operator has an orthonormal basis of eigenvectors. Physicists are well aware, however, of the phenomenon of "continuous spectrum"; thus, when they speak of an "orthonormal basis" they mean either an orthonormal basis in the classic sense or some continuous analog thereof. In the case of the momentum operator P = − i d d x {\textstyle P=-i{\frac {d}{dx}}}  , for example, physicists would say that the eigenvectors are the functions f p ( x ) := e i p x {\displaystyle f_{p}(x):=e^{ipx}}  , which are clearly not in the Hilbert space L 2 ( R ) {\displaystyle L^{2}(\mathbb {R} )}  . (Physicists would say that the eigenvectors are "non-normalizable.") Physicists would then go on to say that these "eigenvectors" are orthonormal in a continuous sense, where the usual Kronecker delta δ i , j {\displaystyle \delta _{i,j}}   is replaced by a Dirac delta function δ ( p − p ′ ) {\displaystyle \delta \left(p-p'\right)}  .

Although these statements may seem disconcerting to mathematicians, they can be made rigorous by use of the Fourier transform, which allows a general L 2 {\displaystyle L^{2}}   function to be expressed as a "superposition" (i.e., integral) of the functions e i p x {\displaystyle e^{ipx}}  , even though these functions are not in L 2 {\displaystyle L^{2}}  . The Fourier transform "diagonalizes" the momentum operator; that is, it converts it into the operator of multiplication by p {\displaystyle p}  , where p {\displaystyle p}   is the variable of the Fourier transform.

The spectral theorem in general can be expressed similarly as the possibility of "diagonalizing" an operator by showing it is unitarily equivalent to a multiplication operator. Other versions of the spectral theorem are similarly intended to capture the idea that a self-adjoint operator can have "eigenvectors" that are not actually in the Hilbert space in question.

Statement of the spectral theorem

Partially defined operators A, B on Hilbert spaces H, K are unitarily equivalent if and only if there is a unitary transformation U : H → K such that

  • U maps dom A bijectively onto dom B,
  • B U ξ = U A ξ , ∀ ξ ∈ dom ⁡ A . {\displaystyle BU\xi =UA\xi ,\qquad \forall \xi \in \operatorname {dom} A.}  

A multiplication operator is defined as follows: Let (X, Σ, μ) be a countably additive measure space and f a real-valued measurable function on X. An operator T of the form

[ T ψ ] ( x ) = f ( x ) ψ ( x ) {\displaystyle [T\psi ](x)=f(x)\psi (x)}  

whose domain is the space of ψ for which the right-hand side above is in L2 is called a multiplication operator.

One version of the spectral theorem can be stated as follows.

Theorem — Any multiplication operator is a (densely defined) self-adjoint operator. Any self-adjoint operator is unitarily equivalent to a multiplication operator.[12]

Other versions of the spectral theorem can be found in the spectral theorem article linked to above.

The spectral theorem for unbounded self-adjoint operators can be proved by reduction to the spectral theorem for unitary (hence bounded) operators.[13] This reduction uses the Cayley transform for self-adjoint operators which is defined in the next section. We might note that if T is multiplication by f, then the spectrum of T is just the essential range of f.

Functional calculus

One important application of the spectral theorem is to define a "functional calculus." That is to say, if h {\displaystyle h}   is a function on the real line and T {\displaystyle T}   is a self-adjoint operator, we wish to define the operator h ( T ) {\displaystyle h(T)}  . If T {\displaystyle T}   has a true orthonormal basis of eigenvectors e j {\displaystyle e_{j}}   with eigenvalues λ j {\displaystyle \lambda _{j}}  , then h ( T ) {\displaystyle h(T)}   is the operator with eigenvectors e j {\displaystyle e_{j}}   and eigenvalues h ( λ j ) {\displaystyle h\left(\lambda _{j}\right)}  . The goal of functional calculus is to extend this idea to the case where T {\displaystyle T}   has continuous spectrum.

Of particular importance in quantum physics is the case in which T {\displaystyle T}   is the Hamiltonian operator H ^ {\displaystyle {\hat {H}}}   and h ( x ) := e − i t x / ℏ {\displaystyle h(x):=e^{-itx/\hbar }}   is an exponential. In this case, the functional calculus should allow us to define the operator

U ( t ) := h ( H ^ ) = e − i t H ^ ℏ , {\displaystyle U(t):=h\left({\hat {H}}\right)=e^{\frac {-it{\hat {H}}}{\hbar }},}  

which is the operator defining the time-evolution in quantum mechanics.

Given the representation of T as the operator of multiplication by f {\displaystyle f}  —as guaranteed by the spectral theorem—it is easy to characterize the functional calculus: If h is a bounded real-valued Borel function on R, then h(T) is the operator of multiplication by the composition h ∘ f {\displaystyle h\circ f}  .

Resolution of the identity

It has been customary to introduce the following notation

E T ⁡ ( λ ) = 1 ( − ∞ , λ ] ( T ) {\displaystyle \operatorname {E} _{T}(\lambda )=\mathbf {1} _{(-\infty ,\lambda ]}(T)}  

where 1 ( − ∞ , λ ] {\displaystyle \mathbf {1} _{(-\infty ,\lambda ]}}   is the characteristic function (indicator function)of the interval ( − ∞ , λ ] {\displaystyle (-\infty ,\lambda ]}  . The family of projection operators ET(λ) is called resolution of the identity for T. Moreover, the following Stieltjes integral representation for T can be proved:

T = ∫ − ∞ + ∞ λ d E T ⁡ ( λ ) . {\displaystyle T=\int _{-\infty }^{+\infty }\lambda d\operatorname {E} _{T}(\lambda ).}  

The definition of the operator integral above can be reduced to that of a scalar valued Stieltjes integral using the weak operator topology. In more modern treatments however, this representation is usually avoided, since most technical problems can be dealt with by the functional calculus.

Formulation in the physics literature

In physics, particularly in quantum mechanics, the spectral theorem is expressed in a way which combines the spectral theorem as stated above and the Borel functional calculus using Dirac notation as follows:

If H is self-adjoint and f is a Borel function,

f ( H ) = ∫ d E | Ψ E ⟩ f ( E ) ⟨ Ψ E | {\displaystyle f(H)=\int dE\left|\Psi _{E}\rangle f(E)\langle \Psi _{E}\right|}  

with

H | Ψ E ⟩ = E | Ψ E ⟩ {\displaystyle H\left|\Psi _{E}\right\rangle =E\left|\Psi _{E}\right\rangle }  

where the integral runs over the whole spectrum of H. The notation suggests that H is diagonalized by the eigenvectors ΨE. Such a notation is purely formal. One can see the similarity between Dirac's notation and the previous section. The resolution of the identity (sometimes called projection valued measures) formally resembles the rank-1 projections | Ψ E ⟩ ⟨ Ψ E | {\displaystyle \left|\Psi _{E}\right\rangle \left\langle \Psi _{E}\right|}  . In the Dirac notation, (projective) measurements are described via eigenvalues and eigenstates, both purely formal objects. As one would expect, this does not survive passage to the resolution of the identity. In the latter formulation, measurements are described using the spectral measure of | Ψ ⟩ {\displaystyle |\Psi \rangle }  , if the system is prepared in | Ψ ⟩ {\displaystyle |\Psi \rangle }   prior to the measurement. Alternatively, if one would like to preserve the notion of eigenstates and make it rigorous, rather than merely formal, one can replace the state space by a suitable rigged Hilbert space.

If f = 1, the theorem is referred to as resolution of unity:

I = ∫ d E | Ψ E ⟩ ⟨ Ψ E | {\displaystyle I=\int dE\left|\Psi _{E}\right\rangle \left\langle \Psi _{E}\right|}  

In the case H eff = H − i Γ {\displaystyle H_{\text{eff}}=H-i\Gamma }   is the sum of an Hermitian H and a skew-Hermitian (see skew-Hermitian matrix) operator − i Γ {\displaystyle -i\Gamma }  , one defines the biorthogonal basis set

H eff ∗ | Ψ E ∗ ⟩ = E ∗ | Ψ E ∗ ⟩ {\displaystyle H_{\text{eff}}^{*}\left|\Psi _{E}^{*}\right\rangle =E^{*}\left|\Psi _{E}^{*}\right\rangle }  

and write the spectral theorem as:

f ( H eff ) = ∫ d E | Ψ E ⟩ f ( E ) ⟨ Ψ E ∗ | {\displaystyle f\left(H_{\text{eff}}\right)=\int dE\left|\Psi _{E}\right\rangle f(E)\left\langle \Psi _{E}^{*}\right|}  

(See Feshbach–Fano partitioning method for the context where such operators appear in scattering theory).

The following question arises in several contexts: if an operator A on the Hilbert space H is symmetric, when does it have self-adjoint extensions? An operator that has a unique self-adjoint extension is said to be essentially self-adjoint; equivalently, an operator is essentially self-adjoint if its closure (the operator whose graph is the closure of the graph of A) is self-adjoint. In general, a symmetric operator could have many self-adjoint extensions or none at all. Thus, we would like a classification of its self-adjoint extensions.

The first basic criterion for essential self-adjointness is the following:[14]

Theorem —  If A is a symmetric operator on H, then A is essentially self-adjoint if and only if the range of the operators A − i {\displaystyle A-i}   and A + i {\displaystyle A+i}   are dense in H.

Equivalently, A is essentially self-adjoint if and only if the operators A ∗ − i {\displaystyle A^{*}-i}   and A ∗ + i {\displaystyle A^{*}+i}   have trivial kernels.[15] That is to say, A fails to be self-adjoint if and only if A ∗ {\displaystyle A^{*}}   has an eigenvector with eigenvalue i {\displaystyle i}   or − i {\displaystyle -i}  .

Another way of looking at the issue is provided by the Cayley transform of a self-adjoint operator and the deficiency indices. (It is often of technical convenience to deal with closed operators. In the symmetric case, the closedness requirement poses no obstacles, since it is known that all symmetric operators are closable.)

Theorem — Suppose A is a symmetric operator. Then there is a unique partially defined linear operator

W ⁡ ( A ) : ran ⁡ ( A + i ) → ran ⁡ ( A − i ) {\displaystyle \operatorname {W} (A):\operatorname {ran} (A+i)\to \operatorname {ran} (A-i)}

 

such that

W ⁡ ( A ) ( A x + i x ) = A x − i x , x ∈ dom ⁡ ( A ) . {\displaystyle \operatorname {W} (A)(Ax+ix)=Ax-ix,\qquad x\in \operatorname {dom} (A).}

 

Here, ran and dom denote the image (in other words, range) and the domain, respectively. W(A) is isometric on its domain. Moreover, the range of 1 − W(A) is dense in H.

Conversely, given any partially defined operator U which is isometric on its domain (which is not necessarily closed) and such that 1 − U is dense, there is a (unique) operator S(U)

S ⁡ ( U ) : ran ⁡ ( 1 − U ) → ran ⁡ ( 1 + U ) {\displaystyle \operatorname {S} (U):\operatorname {ran} (1-U)\to \operatorname {ran} (1+U)}  

such that

S ⁡ ( U ) ( x − U x ) = i ( x + U x ) x ∈ dom ⁡ ( U ) . {\displaystyle \operatorname {S} (U)(x-Ux)=i(x+Ux)\qquad x\in \operatorname {dom} (U).}  

The operator S(U) is densely defined and symmetric.

The mappings W and S are inverses of each other.[clarification needed]

The mapping W is called the Cayley transform. It associates a partially defined isometry to any symmetric densely defined operator. Note that the mappings W and S are monotone: This means that if B is a symmetric operator that extends the densely defined symmetric operator A, then W(B) extends W(A), and similarly for S.

Theorem — A necessary and sufficient condition for A to be self-adjoint is that its Cayley transform W(A) be unitary.

This immediately gives us a necessary and sufficient condition for A to have a self-adjoint extension, as follows:

Theorem — A necessary and sufficient condition for A to have a self-adjoint extension is that W(A) have a unitary extension.

A partially defined isometric operator V on a Hilbert space H has a unique isometric extension to the norm closure of dom(V). A partially defined isometric operator with closed domain is called a partial isometry.

Given a partial isometry V, the deficiency indices of V are defined as the dimension of the orthogonal complements of the domain and range:

n + ( V ) = dim ⁡ dom ⁡ ( V ) ⊥ n − ( V ) = dim ⁡ ran ⁡ ( V ) ⊥ {\displaystyle {\begin{aligned}n_{+}(V)&=\dim \operatorname {dom} (V)^{\perp }\\n_{-}(V)&=\dim \operatorname {ran} (V)^{\perp }\end{aligned}}}  

Theorem — A partial isometry V has a unitary extension if and only if the deficiency indices are identical. Moreover, V has a unique unitary extension if and only if the deficiency indices are both zero.

We see that there is a bijection between symmetric extensions of an operator and isometric extensions of its Cayley transform. The symmetric extension is self-adjoint if and only if the corresponding isometric extension is unitary.

A symmetric operator has a unique self-adjoint extension if and only if both its deficiency indices are zero. Such an operator is said to be essentially self-adjoint. Symmetric operators which are not essentially self-adjoint may still have a canonical self-adjoint extension. Such is the case for non-negative symmetric operators (or more generally, operators which are bounded below). These operators always have a canonically defined Friedrichs extension and for these operators we can define a canonical functional calculus. Many operators that occur in analysis are bounded below (such as the negative of the Laplacian operator), so the issue of essential adjointness for these operators is less critical.

Self-adjoint extensions in quantum mechanics

In quantum mechanics, observables correspond to self-adjoint operators. By Stone's theorem on one-parameter unitary groups, self-adjoint operators are precisely the infinitesimal generators of unitary groups of time evolution operators. However, many physical problems are formulated as a time-evolution equation involving differential operators for which the Hamiltonian is only symmetric. In such cases, either the Hamiltonian is essentially self-adjoint, in which case the physical problem has unique solutions or one attempts to find self-adjoint extensions of the Hamiltonian corresponding to different types of boundary conditions or conditions at infinity.

Example. The one-dimensional Schrödinger operator with the potential V ( x ) = − ( 1 + | x | ) α {\displaystyle V(x)=-(1+|x|)^{\alpha }}  , defined initially on smooth compactly supported functions, is essentially self-adjoint (that is, has a self-adjoint closure) for 0 < α ≤ 2 but not for α > 2. See Berezin and Schubin, pages 55 and 86, or Section 9.10 in Hall.

The failure of essential self-adjointness for α > 2 {\displaystyle \alpha >2}   has a counterpart in the classical dynamics of a particle with potential V ( x ) {\displaystyle V(x)}  : The classical particle escapes to infinity in finite time.[16]

Example. There is no self-adjoint momentum operator p for a particle moving on a half-line. Nevertheless, the Hamiltonian p 2 {\displaystyle p^{2}}   of a "free" particle on a half-line has several self-adjoint extensions corresponding to different types of boundary conditions. Physically, these boundary conditions are related to reflections of the particle at the origin (see Reed and Simon, vol.2).

Suppose A is symmetric densely defined. Then any symmetric extension of A is a restriction of A*. Indeed, A ⊆ B and B symmetric yields B ⊆ A* by applying the definition of dom(A*).

Theorem —  Suppose A is a densely defined symmetric operator. Let

N ± = ran ⁡ ( A ± i ) ⊥ , {\displaystyle N_{\pm }=\operatorname {ran} (A\pm i)^{\perp },}

 

Then

N ± = ker ⁡ ( A ∗ ∓ i ) , {\displaystyle N_{\pm }=\operatorname {ker} (A^{*}\mp i),}

 

and

dom ⁡ ( A ∗ ) = dom ⁡ ( A ¯ ) ⊕ N + ⊕ N − , {\displaystyle \operatorname {dom} \left(A^{*}\right)=\operatorname {dom} \left({\overline {A}}\right)\oplus N_{+}\oplus N_{-},}

 

where the decomposition is orthogonal relative to the graph inner product of dom(A*):

⟨ ξ ∣ η ⟩ graph = ⟨ ξ ∣ η ⟩ + ⟨ A ∗ ξ ∣ A ∗ η ⟩ . {\displaystyle \langle \xi \mid \eta \rangle _{\text{graph}}=\langle \xi \mid \eta \rangle +\left\langle A^{*}\xi \mid A^{*}\eta \right\rangle .}

 

These are referred to as von Neumann's formulas in the Akhiezer and Glazman reference.

We first consider the Hilbert space L 2 [ 0 , 1 ] {\displaystyle L^{2}[0,1]}   and the differential operator

D : ϕ ↦ 1 i ϕ ′ {\displaystyle D:\phi \mapsto {\frac {1}{i}}\phi '}  

defined on the space of continuously differentiable complex-valued functions on [0,1], satisfying the boundary conditions

ϕ ( 0 ) = ϕ ( 1 ) = 0. {\displaystyle \phi (0)=\phi (1)=0.}  

Then D is a symmetric operator as can be shown by integration by parts. The spaces N+, N− (defined below) are given respectively by the distributional solutions to the equation

− i u ′ = i u − i u ′ = − i u {\displaystyle {\begin{aligned}-iu'&=iu\\-iu'&=-iu\end{aligned}}}  

which are in L2[0, 1]. One can show that each one of these solution spaces is 1-dimensional, generated by the functions x → e−x and x → ex respectively. This shows that D is not essentially self-adjoint,[17] but does have self-adjoint extensions. These self-adjoint extensions are parametrized by the space of unitary mappings N+ → N−, which in this case happens to be the unit circle T.

In this case, the failure of essential self-adjointenss is due to an "incorrect" choice of boundary conditions in the definition of the domain of D {\displaystyle D}  . Since D {\displaystyle D}   is a first-order operator, only one boundary condition is needed to ensure that D {\displaystyle D}   is symmetric. If we replaced the boundary conditions given above by the single boundary condition

ϕ ( 0 ) = ϕ ( 1 ) {\displaystyle \phi (0)=\phi (1)}  ,

then D would still be symmetric and would now, in fact, be essentially self-adjoint. This change of boundary conditions gives one particular essentially self-adjoint extension of D. Other essentially self-adjoint extensions come from imposing boundary conditions of the form ϕ ( 1 ) = e i θ ϕ ( 0 ) {\displaystyle \phi (1)=e^{i\theta }\phi (0)}  .

This simple example illustrates a general fact about self-adjoint extensions of symmetric differential operators P on an open set M. They are determined by the unitary maps between the eigenvalue spaces

N ± = { u ∈ L 2 ( M ) : P dist u = ± i u } {\displaystyle N_{\pm }=\left\{u\in L^{2}(M):P_{\operatorname {dist} }u=\pm iu\right\}}  

where Pdist is the distributional extension of P.

Constant-coefficient operators

We next give the example of differential operators with constant coefficients. Let

P ( x → ) = ∑ α c α x α {\displaystyle P\left({\vec {x}}\right)=\sum _{\alpha }c_{\alpha }x^{\alpha }}  

be a polynomial on Rn with real coefficients, where α ranges over a (finite) set of multi-indices. Thus

α = ( α 1 , α 2 , … , α n ) {\displaystyle \alpha =(\alpha _{1},\alpha _{2},\ldots ,\alpha _{n})}  

and

x α = x 1 α 1 x 2 α 2 ⋯ x n α n . {\displaystyle x^{\alpha }=x_{1}^{\alpha _{1}}x_{2}^{\alpha _{2}}\cdots x_{n}^{\alpha _{n}}.}  

We also use the notation

D α = 1 i | α | ∂ x 1 α 1 ∂ x 2 α 2 ⋯ ∂ x n α n . {\displaystyle D^{\alpha }={\frac {1}{i^{|\alpha |}}}\partial _{x_{1}}^{\alpha _{1}}\partial _{x_{2}}^{\alpha _{2}}\cdots \partial _{x_{n}}^{\alpha _{n}}.}  

Then the operator P(D) defined on the space of infinitely differentiable functions of compact support on Rn by

P ( D ) ϕ = ∑ α c α D α ⁡ ϕ {\displaystyle P(\operatorname {D} )\phi =\sum _{\alpha }c_{\alpha }\operatorname {D} ^{\alpha }\phi }  

is essentially self-adjoint on L2(Rn).

Theorem — Let P a polynomial function on Rn with real coefficients, F the Fourier transform considered as a unitary map L2(Rn) → L2(Rn). Then F*P(D)F is essentially self-adjoint and its unique self-adjoint extension is the operator of multiplication by the function P.

More generally, consider linear differential operators acting on infinitely differentiable complex-valued functions of compact support. If M is an open subset of Rn

P ϕ ( x ) = ∑ α a α ( x ) [ D α ϕ ] ( x ) {\displaystyle P\phi (x)=\sum _{\alpha }a_{\alpha }(x)\left[D^{\alpha }\phi \right](x)}  

where aα are (not necessarily constant) infinitely differentiable functions. P is a linear operator

C 0 ∞ ( M ) → C 0 ∞ ( M ) . {\displaystyle C_{0}^{\infty }(M)\to C_{0}^{\infty }(M).}  

Corresponding to P there is another differential operator, the formal adjoint of P

P ∗ f o r m ϕ = ∑ α D α ( a α ¯ ϕ ) {\displaystyle P^{\mathrm {*form} }\phi =\sum _{\alpha }D^{\alpha }\left({\overline {a_{\alpha }}}\phi \right)}  

Theorem — The adjoint P* of P is a restriction of the distributional extension of the formal adjoint to an appropriate subspace of L 2 {\displaystyle L^{2}}  . Specifically:

dom ⁡ P ∗ = { u ∈ L 2 ( M ) : P ∗ f o r m u ∈ L 2 ( M ) } . {\displaystyle \operatorname {dom} P^{*}=\left\{u\in L^{2}(M):P^{\mathrm {*form} }u\in L^{2}(M)\right\}.}

 

The multiplication representation of a self-adjoint operator, though extremely useful, is not a canonical representation. This suggests that it is not easy to extract from this representation a criterion to determine when self-adjoint operators A and B are unitarily equivalent. The finest grained representation which we now discuss involves spectral multiplicity. This circle of results is called the Hahn–Hellinger theory of spectral multiplicity.

Uniform multiplicity

We first define uniform multiplicity:

Definition. A self-adjoint operator A has uniform multiplicity n where n is such that 1 ≤ n ≤ ω if and only if A is unitarily equivalent to the operator Mf of multiplication by the function f(λ) = λ on

L μ 2 ( R , H n ) = { ψ : R → H n : ψ  measurable and  ∫ R ‖ ψ ( t ) ‖ 2 d μ ( t ) < ∞ } {\displaystyle L_{\mu }^{2}\left(\mathbf {R} ,\mathbf {H} _{n}\right)=\left\{\psi :\mathbf {R} \to \mathbf {H} _{n}:\psi {\mbox{ measurable and }}\int _{\mathbf {R} }\|\psi (t)\|^{2}d\mu (t)<\infty \right\}}  

where Hn is a Hilbert space of dimension n. The domain of Mf consists of vector-valued functions ψ on R such that

∫ R | λ | 2   ‖ ψ ( λ ) ‖ 2 d μ ( λ ) < ∞ . {\displaystyle \int _{\mathbf {R} }|\lambda |^{2}\ \|\psi (\lambda )\|^{2}\,d\mu (\lambda )<\infty .}  

Non-negative countably additive measures μ, ν are mutually singular if and only if they are supported on disjoint Borel sets.

Theorem — Let A be a self-adjoint operator on a separable Hilbert space H. Then there is an ω sequence of countably additive finite measures on R (some of which may be identically 0)

{ μ ℓ } 1 ≤ ℓ ≤ ω {\displaystyle \left\{\mu _{\ell }\right\}_{1\leq \ell \leq \omega }}

 

such that the measures are pairwise singular and A is unitarily equivalent to the operator of multiplication by the function f(λ) = λ on

⨁ 1 ≤ ℓ ≤ ω L μ ℓ 2 ( R , H ℓ ) . {\displaystyle \bigoplus _{1\leq \ell \leq \omega }L_{\mu _{\ell }}^{2}\left(\mathbf {R} ,\mathbf {H} _{\ell }\right).}

 

This representation is unique in the following sense: For any two such representations of the same A, the corresponding measures are equivalent in the sense that they have the same sets of measure 0.

Direct integrals

The spectral multiplicity theorem can be reformulated using the language of direct integrals of Hilbert spaces:

Theorem — [18] Any self-adjoint operator on a separable Hilbert space is unitarily equivalent to multiplication by the function λ ↦ λ on

∫ R ⊕ H λ d μ ( λ ) . {\displaystyle \int _{\mathbf {R} }^{\oplus }H_{\lambda }\,d\mu (\lambda ).}

 

Unlike the multiplication-operator version of the spectral theorem, the direct-integral version is unique in the sense that the measure equivalence class of μ (or equivalently its sets of measure 0) is uniquely determined and the measurable function λ ↦ d i m ( H λ ) {\displaystyle \lambda \mapsto \mathrm {dim} (H_{\lambda })}   is determined almost everywhere with respect to μ.[19] The function λ ↦ dim ⁡ ( H λ ) {\displaystyle \lambda \mapsto \operatorname {dim} \left(H_{\lambda }\right)}   is the spectral multiplicity function of the operator.

We may now state the classification result for self-adjoint operators: Two self-adjoint operators are unitarily equivalent if and only if (1) their spectra agree as sets, (2) the measures appearing in their direct-integral representations have the same sets of measure zero, and (3) their spectral multiplicity functions agree almost everywhere with respect to the measure in the direct integral.[20]

Example: structure of the Laplacian

The Laplacian on Rn is the operator

Δ = ∑ i = 1 n ∂ x i 2 . {\displaystyle \Delta =\sum _{i=1}^{n}\partial _{x_{i}}^{2}.}  

As remarked above, the Laplacian is diagonalized by the Fourier transform. Actually it is more natural to consider the negative of the Laplacian −Δ since as an operator it is non-negative; (see elliptic operator).

Theorem — If n = 1, then −Δ has uniform multiplicity mult = 2 {\displaystyle {\text{mult}}=2}  , otherwise −Δ has uniform multiplicity mult = ω {\displaystyle {\text{mult}}=\omega }  . Moreover, the measure μmult may be taken to be Lebesgue measure on [0, ∞).

A self-adjoint operator A on H has pure point spectrum if and only if H has an orthonormal basis {ei}i ∈ I consisting of eigenvectors for A.

Example. The Hamiltonian for the harmonic oscillator has a quadratic potential V, that is

− Δ + | x | 2 . {\displaystyle -\Delta +|x|^{2}.}  

This Hamiltonian has pure point spectrum; this is typical for bound state Hamiltonians in quantum mechanics. As was pointed out in a previous example, a sufficient condition that an unbounded symmetric operator has eigenvectors which form a Hilbert space basis is that it has a compact inverse.

  • Compact operator on Hilbert space
  • Theoretical and experimental justification for the Schrödinger equation
  • Unbounded operator
  • Hermitian adjoint
  • Positive operator

  1. ^ Hall 2013 Corollary 9.9
  2. ^ a b Griffel 2002, p. 238.
  3. ^ a b c d e Griffel 2002, pp. 224–230.
  4. ^ a b Griffel 2002, pp. 240–245.
  5. ^ Hall 2013 Proposition 9.30
  6. ^ Hall 2013 Proposition 9.27
  7. ^ Hall 2013 Proposition 9.28
  8. ^ Hall 2013 Example 9.25
  9. ^ Hall 2013 Theorem 9.41
  10. ^ Berezin & Shubin 1991 p. 85
  11. ^ Hall 2013 Section 9.10
  12. ^ Hall 2013 Theorems 7.20 and 10.10
  13. ^ Hall 2013 Section 10.4
  14. ^ Hall 2013 Theorem 9.21
  15. ^ Hall 2013 Corollary 9.22
  16. ^ Hall 2013 Chapter 2, Exercise 4
  17. ^ Hall 2013 Section 9.6
  18. ^ Hall 2013 Theorems 7.19 and 10.9
  19. ^ Hall 2013 Proposition 7.22
  20. ^ Hall 2013 Proposition 7.24

  • Akhiezer, N. I.; Glazman, I. M. (1981), Theory of Linear Operators in Hilbert Space, Two volumes, Pitman, ISBN 9780486318653
  • Berezin, F. A.; Shubin, M. A. (1991), The Schrödinger Equation, Kluwer
  • Carey, R. W.; Pincus, J. D. (May 1974). "An Invariant for Certain Operator Algebras". Proceedings of the National Academy of Sciences. 71 (5): 1952–1956.
  • Carey, R. W.; Pincus, J. D. (1973). "The structure of intertwining isometries". Indiana University Mathematics Journal. 7 (22): 679--703. doi:10.1512/iumj.1973.22.22056.
  • Griffel, D. H. (2002). Applied functional analysis. Mineola, N.Y: Dover. ISBN 0-486-42258-5. OCLC 49250076.
  • Hall, B. C. (2013), Quantum Theory for Mathematicians, Graduate Texts in Mathematics, vol. 267, Springer, ISBN 978-1461471158
  • Kato, T. (1966), Perturbation Theory for Linear Operators, New York: Springer
  • Moretti, V. (2018), Spectral Theory and Quantum Mechanics:Mathematical Foundations of Quantum Theories, Symmetries and Introduction to the Algebraic Formulation, Springer-Verlag, ISBN 978-3-319-70706-8
  • Narici, Lawrence; Beckenstein, Edward (2011). Topological Vector Spaces. Pure and applied mathematics (Second ed.). Boca Raton, FL: CRC Press. ISBN 978-1584888666. OCLC 144216834.
  • Reed, M.; Simon, B. (1972), Methods of Mathematical Physics, Vol 2, Academic Press
  • Schaefer, Helmut H.; Wolff, Manfred P. (1999). Topological Vector Spaces. GTM. Vol. 8 (Second ed.). New York, NY: Springer New York Imprint Springer. ISBN 978-1-4612-7155-0. OCLC 840278135.
  • Teschl, G. (2009), Mathematical Methods in Quantum Mechanics; With Applications to Schrödinger Operators, Providence: American Mathematical Society
  • Trèves, François (2006) [1967]. Topological Vector Spaces, Distributions and Kernels. Mineola, N.Y.: Dover Publications. ISBN 978-0-486-45352-1. OCLC 853623322.
  • Yosida, K. (1965), Functional Analysis, Academic Press

Retrieved from "https://en.wikipedia.org/w/index.php?title=Self-adjoint_operator&oldid=1101685116"