In linear algebra, every square matrix is associated with a characteristic polynomial. This polynomial encodes several important properties of the matrix, most notably its eigenvalues, its determinant and its trace.
The characteristic polynomial of a graph is the characteristic polynomial of its adjacency matrix. It is a graph invariant, though it is not complete: the smallest pair of non-isomorphic graphs with the same characteristic polynomial have five nodes.1
Given a square matrix A, we want to find a polynomial whose zeros are the eigenvalues of A. For a diagonal matrix A, the characteristic polynomial is easy to define: if the diagonal entries are a1, a2, a3, etc. then the characteristic polynomial will be:
This works because the diagonal entries are also the eigenvalues of this matrix.
For a general matrix A, one can proceed as follows. A scalar λ is an eigenvalue of A if and only if there is an eigenvector v ≠ 0 such that
(where I is the identity matrix). Since v is non-zero, this means that the matrix λ I − A is singular (non-invertible), which in turn means that its determinant is 0. Thus the roots of the function det(λ I − A) are the eigenvalues of A, and it is clear that this determinant is a polynomial in λ.
Some authors define the characteristic polynomial to be det(A - t I). That polynomial differs from the one defined here by a sign (−1)n, so it makes no difference for properties like having as roots the eigenvalues of A; however the current definition always gives a monic polynomial, whereas the alternative definition always has constant term det(A).
Suppose we want to compute the characteristic polynomial of the matrix
We now compute the determinant of
- which is the characteristic polynomial of A.
Its characteristic polynomial is
The polynomial pA(t) is monic (its leading coefficient is 1) and its degree is n. The most important fact about the characteristic polynomial was already mentioned in the motivational paragraph: the eigenvalues of A are precisely the roots of pA(t) (this also holds for the minimal polynomial of A, but its degree may be less than n). The coefficients of the characteristic polynomial are all polynomial expressions in the entries of the matrix. In particular its constant coefficient pA (0) is det(−A) = (−1)n det(A), and the coefficient of t n − 1 is tr(−A) = −tr(A), where tr(A) is the matrix trace of A. (The signs given here correspond to the formal definition given in the previous section;2 for the alternative definition these would instead be det(A) and (−1)n − 1 tr(A) respectively.3) For a 2×2 matrix A, the characteristic polynomial is therefore given by
(Using the language of exterior algebra, one may compactly express the characteristic polynomial of an n×n matrix A as
where tr(ΛkA) is the trace of the kth exterior power of A, with dimension .)
The Cayley–Hamilton theorem states that replacing t by A in the characteristic polynomial (interpreting the resulting powers as matrix powers, and the constant term c as c times the identity matrix) yields the zero matrix. Informally speaking, every matrix satisfies its own characteristic equation. This statement is equivalent to saying that the minimal polynomial of A divides the characteristic polynomial of A.
Two similar matrices have the same characteristic polynomial. The converse however is not true in general: two matrices with the same characteristic polynomial need not be similar.
The matrix A and its transpose have the same characteristic polynomial. A is similar to a triangular matrix if and only if its characteristic polynomial can be completely factored into linear factors over K (the same is true with the minimal polynomial instead of the characteristic polynomial). In this case A is similar to a matrix in Jordan normal form.
If A and B are two square n×n matrices then characteristic polynomials of AB and BA coincide:
More generally, if A is m×n-matrix and B is n×m matrices such that m<n, then AB is m×m and BA is n×n matrix. One has
To prove the first result, recognize that the equation to be proved, as a polynomial in t and in the entries of A and B is a universal polynomial identity. It therefore suffices to check it on an open set of parameter values in the complex numbers. The tuples (A,B,t) where A is an invertible complex n by n matrix, B is any complex n by n matrix, and t is any complex number from an open set in complex space of dimension 2n2 + 1. When A is non-singular our result follows from the fact that AB and BA are similar:
where det is the determinant and I is the identity matrix. The solutions of the characteristic equation are precisely the eigenvalues of the matrix A. The polynomial which results from evaluating the determinant is the characteristic polynomial of the matrix. The term "characteristic equation" is due to Wilhelm Killing.
For example, the matrix
has the characteristic equation
The eigenvalues of this matrix are therefore 20 and 25.
Simple shortcuts exist for low dimension matrices.
For a 3×3 matrix, c2= ½((trA)2−tr(A2)) is the sum of the principal minors of the matrix, and specifies the characteristic polynomial to be
Similarly, for a 4×4 matrix, it evaluates to
Expressions for n×n matrices are increasingly complicated, but tractable, cf. Newton's identities.
The Cayley–Hamilton theorem states that every square matrix satisfies its own characteristic equation.
The term secular function has also been used for what mathematicians now call a characteristic function of a linear operator (in some literature the term secular function is still used). The term comes from the fact that these functions were used to calculate secular perturbations (on a time scale of a century, i.e. slow compared to annual motion) of planetary orbits, according to Lagrange's theory of oscillations.
The characteristic polynomial is defined by the determinant of the matrix with a shift. It has zeros only, without any pole. Commonly, the secular function implies the characteristic polynomial. But, in the strict sense, the secular function has poles as well. Interestingly, the poles are located in the eigenvalues of its sub-matrices. Thus, if the information of the sub-matrices is available, the eigenvalues of the matrix can be described using that kind of information. Furthermore, by partitioning the matrix like matrix tearing or gruing, we can iterate the eigenvalues in a recursive way. According to the methods of partitioning, the variant forms of the secular functions can be built up. However, they are all of the form of a series of the simple rational functions, which have poles at the eigenvalues of the partitioned matrices. For example, we can find a form of secular function in the divide-and-conquer eigenvalue algorithm.
Recently, the secular function has been utilized in signal processing. The estimation problem with uncertainty involves a sort of eigenvalue problem, such as a bounded data uncertainty, total least squares, data least squares, partial least squares, errors-in-variables model, etc. Many cases have been solved using their own secular equations. Some are still trying to find the unique secular equation that can resolve a given uncertainty estimation problem.
As for a numerical aspect, it is known that Newton's method is delicate when finding the roots of the secular equation. The higher-order interpolations are recommended. Among them, a simple rational approximation is a good choice considering the balance between the stability and the computational complexity. It is because the secular equation itself consists of a series of simple rational functions. However, using only interpolation cannot guarantee the stability. Thus fine search algorithms such as bisection steps are still required for accuracy.
Secular equation has several meanings.
In molecular orbital calculations relating to the energy of the electron and its wave function it is also used instead of the characteristic equation.
- "Characteristic Polynomial of a Graph - Wolfram MathWorld". Retrieved August 26, 2011.
- Proposition 28 in these lecture notes
- Theorem 4 in these lecture notes
- "secular equation". Retrieved January 21, 2010.
- T.S. Blyth & E.F. Robertson (1998) Basic Linear Algebra, p 149, Springer ISBN 3-540-76122-5 .
- John B. Fraleigh & Raymond A. Beauregard (1990) Linear Algebra 2nd edition, p 246, Addison-Wesley ISBN 0-201-11949-8 .
- Werner Greub (1974) Linear Algebra 4th edition, pp 120–5, Springer, ISBN 0-387-90110-8 .
- Paul C. Shields (1980) Elementary Linear Algebra 3rd edition, p 274, Worth Publishers ISBN 0-87901-121-1 .
- Gilbert Strang (1988) Linear Algebra and Its Applications 3rd edition, p 246, Brooks/Cole ISBN 0-15-551005-3 .
- R. Skip Garibaldi. The characteristic polynomial and determinant are not ad hoc constructions. http://arxiv.org/abs/math/0203276