Differences between revisions 7 and 8
Revision 7 as of 2025-09-24 18:03:29
Size: 4355
Comment: Simplifying matrix page names
Revision 8 as of 2025-10-16 15:23:31
Size: 4994
Comment: Rewrite
Deletions are marked like this. Additions are marked like this.
Line 13: Line 13:
'''''A'''x'' is a linear transformation: '''''A''''' maps the input vector ''x'' to the output vector ''b''. The square matrix '''''A''''' is a linear transformation mapping vector ''a'' to vector ''b''. There is at least one vector that maps to a linear combination of itself, i.e. multiplying by '''''A''''' works out to multiplying by some scalar ''λ''.
Line 15: Line 15:
For some linear transformations (i.e. some, but not all, matrices '''''A'''''), there are '''certain input vectors''' ''x'' where the output vector ''b'' is also just a linear transformation of the input vector ''x''. In other words, '''''A''''' mapped ''x'' to a scaled version of ''x''. That '''scaling factor''' is notated ''λ''. Each of these vectors is an '''eigenvector'''. Each eigenvector has a corresponding scaling factor ''λ''. These are '''eigenvalues'''.
Line 17: Line 17:
For example, rotation around the ''y'' axis in 3 dimensions by ''θ'' degrees is calculated with: Consider the [[LinearAlgebra/RotationMatrix|matrix representing rotation in the y-axis]]:
Line 27: Line 27:
The the vector ''[0 1 0]]'', and importantly any linear transformation of that unit vector, will not be transformed by this '''''A'''''. (And because the linear transformation involves no scaling, ''λ = 1''.) Given this as '''''A''''', the vector ''[0 1 0]]'' (and any linear transformation of it) maps to itself. The members of this infinite set of vectors are ''all'' eigenvectors of '''''A'''''. (And because there is no scaling, their corresponding eigenvalues ''λ'' are all 1.)
Line 29: Line 29:
The '''certain input vectors''' are the '''eigenvectors''' of '''''A'''''. The '''scaling factors''' are the '''eigenvalues''' of '''''A'''''. Eigenvectors and eigenvalues often include complex numbers.
Line 53: Line 53:
Note that the other eigenvectors and eigenvalues are complex. Note also that the eigenvectors are returned as the eigenvector matrix, usually notated as '''''S'''''. Note that in the above, eigenvectors are returned as an eigenvector matrix. This is usually notated as '''''S'''''.
Line 59: Line 59:
== Definition == == Description ==
Line 63: Line 63:
Unless '''''A''''' is '''defective''', there should be ''n'' unique pairs of eigenvectors and eigenvalues. If there is a repeated eigenvalue, there may not be ''n'' independent eigenvectors. Given a matrix of size ''n x n'', either there are ''n'' unique pairs of eigenvectors and eigenvalues, or the matrix is '''defective'''.
Line 69: Line 69:
Only square matrices have eigenvectors.
Line 71: Line 73:
The sum of the eigenvalues is the trace (sum of diagonal). The product of the eigenvalues is the [[LinearAlgebra/Determinant|determinant]]. The [[LinearAlgebra/Trace|trace]] is the sum of eigenvalues. The [[LinearAlgebra/Determinant|determinant]] is the product the eigenvalues.

A [[LinearAlgebra/Diagonalization|diagonal matrix]] is a trivial case because...
 * the columns are its eigenvectors
 * the numbers in the diagonal are its eigenvalues

This also means that any diagonalizable matrix of size ''n x n'' must have ''n'' unique pairs of eigenvectors and eigenvalues, and cannot be defective.
Line 77: Line 85:
== Solution == == Calculation ==
Line 79: Line 87:


=== Special Cases ===

A [[LinearAlgebra/Diagonalization|diagonal matrix]] is a special and trivial case for finding eigenvalues and eigenvectors.



=== Finding Eigenvalues ===
=== Conventional Method ===
Line 96: Line 96:
In a simple 2 x 2 matrix, this looks like: In a simple ''2 x 2'' matrix, this looks like:
Line 114: Line 114:
This leads to the '''characteristic polynomial''' of '''''A'''''; solving for the roots, as through either factorization or the quadratic formula, gives the eigenvalues.
Line 115: Line 116:
Because eigenvectors are characterized by...
Line 116: Line 118:
=== Finding Eigenvectors === '''''A'''x = λx''
Line 118: Line 120:
Because eigenvectors are characterered by: '''''A'''x - λx = 0''
Line 120: Line 122:
 * '''''A'''x = λx''
 * which can be rewritten as '''''A'''x - λx = 0''
 * which can be refactored as ''('''A''' - λ'''I''')x = 0''
''('''A''' - λ'''I''')x = 0''
Line 124: Line 124:
Eigenvectors can be solved for given eigenvalues. For each given ''λ,,i,,'', substitute it into that final equation and solve for the corresponding ''x,,i,,''. ...eigenvectors can be solved for given eigenvalues using substitution.
Line 126: Line 126:
In a simple 2 x 2 matrix, this looks like: In a simple ''2 x 2'' matrix, this looks like:
Line 152: Line 152:
=== Shortcut ===

This method is only applicable to ''2 x 2'' matrices.

 * Because the trace is equal to the sum of eigenvalues, it follows that 1/2 of the trace is also the mean of the eigenvalues.
 * Because the characteristic polynomial must be quadratic, the eigenvalues must be evenly spaced from the center (i.e., the mean). Given the above mean as ''m'' and an unknown distance as ''d'', the eigenvalues must be ''(m-d)'' and ''(m+d)''.
 * By definition of the determinant, ''|'''A'''| = (m-d)(m+d) = m^2^ - d^2^''. This can be solved for ''d''.

Altogether,

{{attachment:shortcut.svg}}


Eigenvalues and Eigenvectors

Matrices are characterized by their eigenvalues and eigenvectors.


Introduction

The square matrix A is a linear transformation mapping vector a to vector b. There is at least one vector that maps to a linear combination of itself, i.e. multiplying by A works out to multiplying by some scalar λ.

Each of these vectors is an eigenvector. Each eigenvector has a corresponding scaling factor λ. These are eigenvalues.

Consider the matrix representing rotation in the y-axis:

┌                 ┐
│  cos(θ) 0 sin(θ)│
│       0 1      0│
│ -sin(θ) 0 cos(θ)|
└                 ┘

Given this as A, the vector [0 1 0]] (and any linear transformation of it) maps to itself. The members of this infinite set of vectors are all eigenvectors of A. (And because there is no scaling, their corresponding eigenvalues λ are all 1.)

Eigenvectors and eigenvalues often include complex numbers.

julia> using LinearAlgebra

julia> A = [0 0 1; 0 1 0; -1 0 0]
3×3 Matrix{Int64}:
  0  0  1
  0  1  0
 -1  0  0

julia> eigvals(A)
3-element Vector{ComplexF64}:
 0.0 - 1.0im
 0.0 + 1.0im
 1.0 + 0.0im

julia> eigvecs(A)
3×3 Matrix{ComplexF64}:
 0.707107-0.0im       0.707107+0.0im       0.0+0.0im
      0.0-0.0im            0.0+0.0im       1.0+0.0im
      0.0-0.707107im       0.0+0.707107im  0.0+0.0im

Note that in the above, eigenvectors are returned as an eigenvector matrix. This is usually notated as S.


Description

Eigenvalues and eigenvectors are the pairs of λ and x that satisfy Ax = λx and |A - λI| = 0.

Given a matrix of size n x n, either there are n unique pairs of eigenvectors and eigenvalues, or the matrix is defective.

Properties

Only square matrices have eigenvectors.

Adding nI to A does not change its eigenvectors and adds n to the eigenvalues.

The trace is the sum of eigenvalues. The determinant is the product the eigenvalues.

A diagonal matrix is a trivial case because...

  • the columns are its eigenvectors
  • the numbers in the diagonal are its eigenvalues

This also means that any diagonalizable matrix of size n x n must have n unique pairs of eigenvectors and eigenvalues, and cannot be defective.


Calculation

Conventional Method

Because eigenvalues are characterized by |A - λI| = 0, they can be solved for by:

  • subtracting λ from each value on the diagonal

  • formulating the determinant for this difference
  • setting the formulation for 0
  • solving for λ

In a simple 2 x 2 matrix, this looks like:

|    A   -  λI   | = 0

│ ┌    ┐  ┌    ┐ │
│ │ a b│ -│ λ 0│ │ = 0
│ │ c d│  │ 0 λ│ │
│ └    ┘  └    ┘ │

│ ┌        ┐ │
│ │ a-λ   b│ │ = 0
│ │   c d-λ│ │
│ └        ┘ │

(a-λ)(d-λ) - bc = 0

This leads to the characteristic polynomial of A; solving for the roots, as through either factorization or the quadratic formula, gives the eigenvalues.

Because eigenvectors are characterized by...

Ax = λx

Ax - λx = 0

(A - λI)x = 0

...eigenvectors can be solved for given eigenvalues using substitution.

In a simple 2 x 2 matrix, this looks like:

(   A    -  λI   )   x  =   0

/ ┌    ┐  ┌    ┐ \ ┌  ┐   ┌  ┐
│ │ a b│ -│ λ 0│ │ │ u│ = │ 0│
│ │ c d│  │ 0 λ│ │ │ v│   │ 0│
\ └    ┘  └    ┘ / └  ┘   └  ┘

┌        ┐ ┌  ┐   ┌  ┐
│ a-λ   b│ │ u│ = │ 0│
│   c d-λ│ │ v│   │ 0│
└        ┘ └  ┘   └  ┘

┌        ┐ ┌  ┐   ┌  ┐
│ a-λ   b│ │ u│ = │ 0│
│   c d-λ│ │ v│   │ 0│
└        ┘ └  ┘   └  ┘

(a-λ)u + (b)v = 0
(c)u + (d-λ)v = 0

Shortcut

This method is only applicable to 2 x 2 matrices.

  • Because the trace is equal to the sum of eigenvalues, it follows that 1/2 of the trace is also the mean of the eigenvalues.
  • Because the characteristic polynomial must be quadratic, the eigenvalues must be evenly spaced from the center (i.e., the mean). Given the above mean as m and an unknown distance as d, the eigenvalues must be (m-d) and (m+d).

  • By definition of the determinant, |A| = (m-d)(m+d) = m2 - d2. This can be solved for d.

Altogether,

shortcut.svg


CategoryRicottone

LinearAlgebra/EigenvaluesAndEigenvectors (last edited 2026-02-03 23:52:07 by DominicRicottone)