eigenvalue is the scalar needed to scale the basis element of a one dimensional invariant subspace of a Linear Map to represent the behavior of the map:
Note we require v \neq 0 because otherwise all scalars count. eigenvector is a vector that forms the basis list of length 1 of that 1-D invariant subspace under T. “operators own eigenvalues, eigenvalues own eigenvectors” Why is eigenvalue consistent per eigenvector? Because a linear map has to act on the same way to something’s basis as it does to the whole space. Motivation Take some subspace U \subset V:
Now, if T|_{U} is an operator on U, U would be an invariant subspace of T of dimension 1 (its basis being the list \{v\}). Therefore, for some vector v \in U (basically like various scalings of v), T will always send back to U so we can represent it yet again with another scalar on v, like \lambda v. In this case, then, we can write that:
And then the usual definition of eigenvalues persist. constituents linear map T \in \mathcal{L}(V) vector v \in V, such that v \neq 0 scalar \lambda \in \mathbb{F} requirements If there exists v \in V such that v\neq 0 and:
then, \lambda is called an eigenvalue, and v the eigenvector. additional information properties of eigenvalues Suppose V in finite-dimensional, T \in \mathcal{L}(V) and \lambda \in \mathbb{F}, then: \lambda is an eigenvalue of T T - \lambda I is not injective T - \lambda I is not surjective T - \lambda I is not invertable Showing one shows all. Proof: 1 \implies 2 Suppose \lambda is an eigenvalue of T. Then, we have some v \in V such that:
Now:
the last step by (T+S)v = Tv+Sv, the property of the vector space of \mathcal{L}(V) (or any \mathcal{L}). And therefore, v \in null\ (T-\lambda I), and v\neq 0. And so null\ (T-\lambda I) \neq \{0\} and so T-\lambda I is not injective, as desired. The reverse of this result shows the opposite direction that 1 \implies 2. The others I \in \mathcal{L}(V), T \in \mathcal{L}(V), \mathcal{L}(V) is closed, so (T - \lambda I) \in \mathcal{L}(V), and so it is an operator. Having 2) implies all other conditions of non-injectivity, non-surjectivity, non-invertiblility by injectivity is surjectivity in finite-dimensional operators list of eigenvectors are linearly independent Let T \in \mathcal{L}(V), suppose \lambda_{j} are distinct eigenvalues of T, and v_1, \ldots, v_{m} the corresponding eigenvectors, then v_1, \ldots, v_{m} is linearly independent. proof: We will show this by contradiction. Suppose v_1, \ldots, v_{m} are linearly dependent; then, by the Linear Dependence Lemma, \exists v_{j} such that:
Meaning:
Given the list is a list of eigenvalues, we can apply T to both sides to get:
We can also get another definition for \lambda_{j} v_{j} by simply multiplying the definition for v_{j} above by \lambda_{j}:
Now, subtracting our two definitions of \lambda_{j} v_{j}, we get:
Recall now that the eigenvalue list \lambda_{j} are distinct. This means all \lambda_{j} - \lambda_{k \neq j} \neq 0. No v_{j} =0; so if we choose the smallest positive integer for j, the list before it v_1, \dots, v_{j-1} is linearly independent (as no value in that list would satisfy the Linear Dependence Lemma). This makes a_{j} =\dots =a_{j-1} = 0. And yet, substituting this back into the expression for v_{j}, we have v_{j} = 0, reaching contradiction. So therefore, the list of eigenvectors are linearly independent. \blacksquare operators on finite dimensional V has at most dim V eigenvalues As a corollary of the above result, suppose V is finite dimensional; then, each operator on V has at most dim\ V distinct eigenvalues because their eigenvectors form an linearly independent list and length of linearly-independent list \leq length of spanning list. eigenspaces are disjoint the eigenspaces of a Linear Map form a direct sum: proof: Corollary of result above. Because eigenvectors (i.e. bases) from distinct eigenspaces are linearly independent. So the only way to write 0 is by taking each to 0. So by taking the bases all to 0, you take the 0 vector from each space, which shows that the eigenspaces are a direct sum. \blacksquare finding eigenvalues with actual numbers \begin{equation} \lambda_{j} \in Spec(T) \Rightarrow det(\lambda_{j}I-T) = 0 \end{equation} The right polynomial det(\lambda_{j} I-T) = 0 is named the “characteristic polynomial.” natural choordinates of a map Given the eigenvectors (x+,y+), (x-,y-), we can change coordinates of your matrix into the natural choordinates.
This makes scaling matricides much much easier. If you think about multiplying the above matrix n times, the inverse and non-inverse cancells out. similar matrices Let A,B be defined:
and of course:
where, A,B,C \in \mathcal{L}(V) A, B has the same eigenvalues. invertable matricies Let T \in \mathcal{L}(V) be invertable. If \lambda is an eigenvalue of T, then \frac{1}{\lambda} is an eigenvalue of T. Furthermore, T and T^{-1} share eigenvectors with eigenvalues \lambda and \frac{1}{\lambda} symmetric matricies have a real basis of eigenvalues this falls out of the real spectral theorem.