In mathematics, a Hermitian matrix is a square matrix with complex that is equal to its own conjugate transpose - that is, if the element in the ith row and jth column is equal to the complex conjugate of the element in the jth row and ith column, for all indices i and j:
Here is an example of a Hermitian matrix:
Every Hermitian matrix is normal, and the finite-dimensional spectral theorem applies. It says that any Hermitian matrix can be diagonalized by a unitary matrix, and that the resulting diagonal matrix has only real entries. This means that all eigenvaluess of a Hermitian matrix are real, and, moreover, eigenvectors with distinct eigenvalues are orthogonal. It is possible to find an orthonormal basis of Cn consisting only of eigenvectors.
If the eigenvalues of a Hermitian matrix are all positive, then the matrix is positive definite.
for all elements x and y of H. Here, the parentheses denote the inner product given on H.
This definition agrees with the one given above if we take as H the Hilbert space Cn with the standard dot product and interpret a square matrix as a linear operator on this Hilbert space. It is however much more general as there are important infinite-dimensional Hilbert spaces.
The spectrum of any Hermitian operator is real; in particular all its eigenvalues are real. A version of the spectral theorem also applies to Hermitian operators; while the eigenvectors to different eigenvalues are orthogonal, in general it is not true that the Hilbert space H admits an orthonormal basis consisting only of eigenvectors of the operator. In fact, Hermitian operators need not have any eigenvalues or eigenvectors at all.
In the mathematical formulation of quantum mechanics, one considers even more general Hermitian operators: they are only defined on a dense subspace of a Hilbert space and don't have to be continuous.
For example, consider the complex Hilbert space L2[0,1] and the differential operator A = d2 / dx2, defined on the subspace consisting of all differentiable functions f : [0,1] → C with f(0) = f(1) = 0. Then integration by parts easily proves that A is Hermitian. Its eigenfunctions are the sinusoids sin(nπx) for n = 1,2,..., with the real eigenvalues n2π2; the well-known orthogonality of the sine functions follows as a consequence of the Hermitian property.
Another example: the complex Hilbert space L2(R), and the operator which multiplies a given function by x: