Hermitian matrix

In mathematics, a Hermitian matrix (or self-adjoint matrix) is a square matrix with complex entries that is equal to its own conjugate transpose—that is, the element in the i-th row and j-th column is equal to the complex conjugate of the element in the j-th row and i-th column, for all indices i and j:

a_{ij} = \overline{a_{ji}} or A = \overline {A^\text{T}}, in matrix form.

Hermitian matrices can be understood as the complex extension of real symmetric matrices.

If the conjugate transpose of a matrix A is denoted by A^\dagger, then the Hermitian property can be written concisely as

 A = A^\dagger.

Hermitian matrices are named after Charles Hermite, who demonstrated in 1855 that matrices of this form share a property with real symmetric matrices of always having real eigenvalues.

Examples

See the following example:


\begin{bmatrix}
2 & 2+i & 4 \\
2-i & 3 & i \\
4 & -i & 1 \\
\end{bmatrix}

The diagonal elements must be real, as they must be their own complex conjugate.

Well-known families of Pauli matrices, Gell-Mann matrices and their generalizations are Hermitian. In theoretical physics such Hermitian matrices are often multiplied by imaginary coefficients,[1][2] which results in skew-Hermitian matrices (see below).

Here we offer another useful Hermitian matrix using an abstract example. If a square matrix  A equals the multiplication of a matrix and its conjugate transpose, that is,  A=BB^\dagger , then  A is a Hermitian positive semi-definite matrix. Furthermore, if  B is row full-rank, then  A is positive definite.

Properties

\; E_{jj} for 1\leq j\leq n (n matrices)
together with the set of matrices of the form
\; E_{jk}+E_{kj} for 1\leq j<k\leq n (n2n/2 matrices)
and the matrices
\; i(E_{jk}-E_{kj}) for 1\leq j<k\leq n (n2n/2 matrices)
where i denotes the complex number \sqrt{-1}, known as the imaginary unit.
 A = \sum _j \lambda_j u_j u_j ^\dagger ,
where \lambda_j are the eigenvalues on the diagonal of the diagonal matrix \; \Lambda .

Further properties

Additional facts related to Hermitian matrices include:

C = A+B \quad\mbox{with}\quad A = \frac{1}{2}(C + C^{\dagger}) \quad\mbox{and}\quad B = \frac{1}{2}(C - C^{\dagger}).
Proof:  \det(A) = \det(A^\mathrm{T})\quad \Rightarrow \quad \det(A^\dagger) = \det(A)^*
Therefore if A=A^\dagger\quad \Rightarrow \quad \det(A) = \det(A)^*.
(Alternatively, the determinant is the product of the matrix's eigenvalues, and as mentioned before, the eigenvalues of a Hermitian matrix are real.)

Rayleigh quotient

Main article: Rayleigh quotient

In mathematics, for a given complex Hermitian matrix M and nonzero vector x, the Rayleigh quotient[3] R(M, x), is defined as:[4][5]


R(M,x) := {x^{*} M x \over x^{*} x}.

For real matrices and vectors, the condition of being Hermitian reduces to that of being symmetric, and the conjugate transpose x^{*} to the usual transpose x'. Note that R(M, c x) = R(M,x) for any non-zero real scalar c. Recall that a Hermitian (or real symmetric) matrix has real eigenvalues. It can be shown that, for a given matrix, the Rayleigh quotient reaches its minimum value \lambda_\min (the smallest eigenvalue of M) when x is v_\min (the corresponding eigenvector). Similarly, R(M, x) \leq \lambda_\max and R(M, v_\max) = \lambda_\max.

The Rayleigh quotient is used in the min-max theorem to get exact values of all eigenvalues. It is also used in eigenvalue algorithms to obtain an eigenvalue approximation from an eigenvector approximation. Specifically, this is the basis for Rayleigh quotient iteration.

The range of the Rayleigh quotient (for matrix that is not necessarily Hermitian) is called a numerical range, (or spectrum in functional analysis). When the matrix is Hermitian, the numerical range is equal to the spectral norm. Still in functional analysis, \lambda_\max is known as the spectral radius. In the context of C*-algebras or algebraic quantum mechanics, the function that to M associates the Rayleigh-Ritz quotient R(M,x) for a fixed x and M varying through the algebra would be referred to as "vector state" of the algebra.

See also

References

  1. Frankel, Theodore (2004). The geometry of physics: an introduction. Cambridge University Press. p. 652. ISBN 0-521-53927-7.
  2. Physics 125 Course Notes at California Institute of Technology
  3. Also known as the Rayleigh–Ritz ratio; named after Walther Ritz and Lord Rayleigh.
  4. Horn, R. A. and C. A. Johnson. 1985. Matrix Analysis. Cambridge University Press. pp. 176180.
  5. Parlet B. N. The symmetric eigenvalue problem, SIAM, Classics in Applied Mathematics,1998

External links

This article is issued from Wikipedia - version of the Wednesday, March 09, 2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.