Eigencircles are a concept I came across on this math stackexchange forum. They are a visualization of 2x2 matrices that provide a fascinating geometric understanding of matrix properties and their connection to eigenvectors and eigenvalues. I aim to summarize my understanding of eigencircles in this article.
What are Eigencircles?
An eigenvalue of a square matrix A is a number λ such that Ax=λx for any non-zero vector x. Let A=[acbd] and x=(xy). Then we can write this as
[acbd](xy)=[λ00λ](xy)
Now in the original article, Englefield and Farr state that matrices of the form [λ00λ] are isomorphic to the reals under matrix addition and multiplication. They then utilize the field isomorphism
[λ−μμλ]→λ+iμ
to rewrite (1) as
[acbd](xy)=[λ−μμλ](xy)
After skimming through a few articles on abstract algebra, this made sense. But I came across another approach to arrive at (2) that felt more intuitive. I describe this next.
From the SVD decomposition of a matrix, we know that every linear transformation represented by the matrix A can be written as A=PΣQT where P,Q are othgonal matrices and Σ is a diagonal matrix. For the 2x2 case, the only orthogonal matrices are rotation and refelction matrices while the diagonal matrix is just a scaling matrix. Therefore, the 2x2 matrix A simply rotates and scales every vector it is applied to. Notice that reflecting a vector along some axis of rotation can be expressed as a rotation.
In other words, for any vector x, Ax results in a rotation by angle θ=∠(x,Ax) and a scaling by factor s. More formally, we can write this as
∀x=[xy],∃(s,θ):Ax=s⋅[cosθ+sinθ−sinθcosθ][xy]
We can rewrite the matrix on the right hand side of the equality to be more succint
s⋅[cosθ+sinθ−sinθcosθ]=[λμ−μλ]
Now you may notice that this deviates from (2) in that the minus signs on μ are reversed. This is the same as reversing the positive direction of the angles. Notice that
In Figure 1, the left diagram corresponds to the case where we take [λμ−μλ] and the right corresponds to the case where we consider [λ−μμλ]. Therefore, it is equivalant to continue using either of the derivations. I will continue using (2).
Following the general procedure to determine eigenvalues but using the matrix specified in the RHS of (2), we get
a−λc+μb−μd−λ=0
This results in the following equation:
(a−λ)(d−λ)−(b−μ)(c+μ)=0
or, equivalently
λ2+μ2−(a+d)λ−(b−c)μ+ad−bc=0
Notice that ad−bc is just det(A). Now we define
f=2a+dg=2b−cr2=f2+g2
Using (5),(6),(7) to simplify (4) we get
λ2−2fλ+f2+μ2−2gμ+g2=r2−det(A)
which is
(λ−f)2+(μ−g)2=ρ2
where ρ2=r2−det(A). (8) gives the equation for the eigencircle of A.
Deriving matrix properties from Eigencircles
Okay, now consider (3) again. We know that the determinant of any matrix with a zero row or a zero column is zero. Considering the values of λ,μ that would produce zero rows and columns, we get the following points: (a,b),(d,b),(a,−c),(d,−c). These points must be on the eigencircle since they satisfy (3). Moreover, since the center of our eigencircle is defined to be the midpoint of (a,d) and (b,−c)
(f,g)=(2a+d,2b−c)
these points define a rectangle.
Defining the points F=(a,b),H=(a,−c),E=(d,−c),G=(d,b), we get the following graph.
Before moving forward, you may notice the axes on Figure 2 corresponding to x,y - these diagrams are from the original article and I was too lazy to refactor them. For the purposes of this article, x=λ and y=μ.
This visualization has some very interesting properties. Firstly, notice that if (λ,0) lies on the circle, then λ is an eigenvalue of A. Let L=(λ,0), then the vector LE is the corresponding eigenvector for λ.
To see why, note that the vector LE=(d−c)−(λ0)=(d−λ−c). If this is an eigenvector corresponding to eigenvalue λ, then it must be in the nullspace of the matrix [a−λcbd−λ].
Let’s see if this is the case:
[a−λcbd−λ](d−λ−c)=((a−λ)(d−λ)−bc0)
But we know that (a−λ)(d−λ)−bc=0 from (3) (since (λ,0) lies on the eigencircle). Therefore, LE is indeed an eigenvector correspoinding to λ.
We can also use the eigencircle to prove that the product of eigenvalues for a 2x2 matrix A is equal to the value of its determinant. We use the “power of a point” theorem to show this. The power of any point P given by Π(P) with respect to some circle c with center O is defined as
Π(P)=∣PO∣2−r2
Geometrically, the power of a point states that “[if some] line through any point P meets a circle in points Q and R, the product of the lengths, PQ and PR, is the same for any direction of the line. The lengths are signed, so the product is positive (negative) when P is outside (inside)
the circle.”
Consider Figure 4. What the power of a point theorem says is that the (signed) product of the lengths PX⋅PX′=PY⋅PY′=Π(P). For more on the power of a point, refer to this link.
Consider (9). For the eigencircle, we can rewrite this as (Px−f)2+(Py−g)2−ρ2. Notice that this expression can be expressed as the LHS of (3) by construction. So we get that the power of a point with respect to a eigencircle is given by
Π(P)=(a−Px)(d−Px)−(b−Py)(c+Py)
If we choose P=O=(0,0), we get that Π(O)=det(A)! From Figure 3, we know that the (signed) lengths of the line segments OL1 and OL2 specify the eigenvalues. Well, from the power of a point theorem, we know that OL1⋅OL2=Π(O)=det(A). Personally, I found this fascinating.
Finally, I am going to describe how the eigencircle relates to complex eigenvalues. Consider Figure 5a.
From the power of a point theorem, we know that the power of Y is given by
Π(Y)=YL1⋅YL2=YM⋅YN
As YL1=−YL2, we have that YL12=YM⋅YN>0 (remember that we consider signed lengths). Knowing this, we can express the eigenvalues (L1,L2) in the following form
λ=OY±−YM⋅YN
Now this is also valid for cases where the x-axis does not intersect the eigencircles (we will see why in a minute). Consider Figure 5b. We see that the quantity YM⋅YN is now positive, there the value inside the square in (11) is negative, giving us a complex value. If YV is a tangent to the circle at V, then, by the power of a point theorem, we have that YM⋅YN=YV2. Hence, we get the complex eigenvalues
λ=OY±iYV
Okay, so why is (11) true when the x-axis does not intersect the eigencircle? Let’s figure out the coordinates Y,M,N. Well, we can see that Y=(f,0),M=(f,g−ρ),N=(f,g+ρ) (recall definitions 5,6,7). Plugging this into (11) we get
λ=f±−(g2−ρ2)
or equivalently,
λ=f±ρ2−g2
Notice that if the x-axis does not intersect the eigencircle, then we have that ρ2<g2 and therefore ρ2−g2<0, just as we had derived earlier. For this case, we need to realize that ρ2−g2 is undefined and we need to factor out the −1. More on this below.
For λ to be an eigenvalue, we know that equation (3) must be satisfied for (λ,0). Notice that equation (3) can be rewritten as
(λ−f)2+(μ−g)2−ρ2=0
This is just (8) with the ρ2 moved to the LHS. Plugging in (λ,0), we get
as required. Therefore, you can determine all (both real and complex) eigenvalues using the eigencircle.
Visualizing the eigenvectors corresponding to complex eigenvalues would require an additional axis to account for the imaginary value. Let λ=f±ih where h=YV. Consider Figure 6.
LE
and
KE
represent the complex eigenvectors. Notice
LE=OE−OL=d−c0−f0h
represents the complex eigenvector (d−(f+ih)−c)=(d−f−ih−c).
I am going to stop there. It was fun reading about eigencircles. I believe the way in which the properties of eigenvectors and eigenvalues expressed themselves through geometry was beautiful. Refer to [1] and [2] for the original articles on eigencircles.
References
Englefield, M. J., & Farr, G. E. (2006). Eigencircles of 2 x 2 Matrices. Mathematics Magazine Vol. 79 Oct.,2006, 281-289.
Englefield, M. J., & Farr, G. E. (2010). Eigencircles and associated surfaces. The Mathematical Gazette Vol.94 No. 531 (November 2010), 438-449.