skip to content

Eigencircles

/ 10 min read

Introduction

Eigencircles are a concept I came across on this math stackexchange forum. They are a visualization of 2x2 matrices that provide a fascinating geometric understanding of matrix properties and their connection to eigenvectors and eigenvalues. I aim to summarize my understanding of eigencircles in this article.

What are Eigencircles?

An eigenvalue of a square matrix AA is a number λ\lambda such that Ax=λxAx = \lambda x for any non-zero vector xx. Let A=[abcd]A = \begin{bmatrix} a & b \\ c & d \end{bmatrix} and x=(xy)x = \begin{pmatrix} x \\ y \end{pmatrix}. Then we can write this as

[abcd](xy)=[λ00λ](xy)\begin{equation} \begin{bmatrix} a & b \\ c & d \end{bmatrix} \begin{pmatrix} x \\ y \end{pmatrix} = \begin{bmatrix} \lambda & 0 \\ 0 & \lambda \end{bmatrix} \begin{pmatrix} x \\ y \end{pmatrix} \end{equation}

Now in the original article, Englefield and Farr state that matrices of the form [λ00λ]\begin{bmatrix} \lambda & 0 \\ 0 & \lambda \end{bmatrix} are isomorphic to the reals under matrix addition and multiplication. They then utilize the field isomorphism

[λμμλ]λ+iμ\begin{bmatrix} \lambda & \mu \\ -\mu & \lambda \end{bmatrix} \rightarrow \lambda + i\mu

to rewrite (1)(1) as

[abcd](xy)=[λμμλ](xy)\begin{equation} \begin{bmatrix} a & b \\ c & d \end{bmatrix} \begin{pmatrix} x \\ y \end{pmatrix} = \begin{bmatrix} \lambda & \mu \\ -\mu & \lambda \end{bmatrix} \begin{pmatrix} x \\ y \end{pmatrix} \end{equation}

After skimming through a few articles on abstract algebra, this made sense. But I came across another approach to arrive at (2)(2) that felt more intuitive. I describe this next.

From the SVD decomposition of a matrix, we know that every linear transformation represented by the matrix AA can be written as A=PΣQTA = P\Sigma Q^{T} where P,QP, Q are othgonal matrices and Σ\Sigma is a diagonal matrix. For the 2x2 case, the only orthogonal matrices are rotation and refelction matrices while the diagonal matrix is just a scaling matrix. Therefore, the 2x2 matrix AA simply rotates and scales every vector it is applied to. Notice that reflecting a vector along some axis of rotation can be expressed as a rotation.

In other words, for any vector xx, AxAx results in a rotation by angle θ=(x,Ax)\theta = \angle (x, Ax) and a scaling by factor ss. More formally, we can write this as

x=[xy],(s,θ):Ax=s[cosθsinθ+sinθcosθ][xy]\forall x = \begin{bmatrix} x \\ y \end{bmatrix}, \exists (s, \theta) : Ax = s \cdot \begin{bmatrix} \cos\theta & -\sin\theta \\ +\sin\theta & \cos\theta \end{bmatrix} \begin{bmatrix} x \\ y \end{bmatrix}

We can rewrite the matrix on the right hand side of the equality to be more succint

s[cosθsinθ+sinθcosθ]=[λμμλ]s \cdot \begin{bmatrix} \cos\theta & -\sin\theta \\ +\sin\theta & \cos\theta \end{bmatrix} = \begin{bmatrix} \lambda & -\mu \\ \mu & \lambda \end{bmatrix}

Now you may notice that this deviates from (2)(2) in that the minus signs on μ\mu are reversed. This is the same as reversing the positive direction of the angles. Notice that

s[cos(θ)sin(θ)+sin(θ)cos(θ)]=s[cosθ+sinθsinθcosθ]=[λμμλ]s \cdot \begin{bmatrix} \cos(-\theta) & -\sin(-\theta) \\ +\sin(-\theta) & \cos(-\theta) \end{bmatrix} = s \cdot \begin{bmatrix} \cos\theta & +\sin\theta \\ -\sin\theta & \cos\theta \end{bmatrix} = \begin{bmatrix} \lambda & \mu \\ -\mu & \lambda \end{bmatrix}
eigencircle

Figure 1: Taken from https://eigencircles.heavisidesdinner.com/eigenC/eigenC_book.html

In Figure 1, the left diagram corresponds to the case where we take [λμμλ]\begin{bmatrix} \lambda & -\mu \\ \mu & \lambda \end{bmatrix} and the right corresponds to the case where we consider [λμμλ]\begin{bmatrix} \lambda & \mu \\ -\mu & \lambda \end{bmatrix}. Therefore, it is equivalant to continue using either of the derivations. I will continue using (2)(2).

Following the general procedure to determine eigenvalues but using the matrix specified in the RHS of (2), we get

aλbμc+μdλ=0\begin{vmatrix} a - \lambda & b-\mu \\ c + \mu & d - \lambda \end{vmatrix} = 0

This results in the following equation:

(aλ)(dλ)(bμ)(c+μ)=0\begin{equation} (a-\lambda)(d-\lambda)-(b-\mu)(c+\mu) = 0 \end{equation}

or, equivalently

λ2+μ2(a+d)λ(bc)μ+adbc=0\begin{equation} \lambda^{2} + \mu^{2} - (a+d)\lambda - (b-c)\mu + ad - bc = 0 \end{equation}

Notice that adbcad - bc is just det(A)\det(A). Now we define

f=a+d2\begin{equation} f = \frac{a+d}{2} \end{equation} g=bc2\begin{equation} g = \frac{b-c}{2} \end{equation} r2=f2+g2\begin{equation} r^{2} = f^{2} + g^{2} \end{equation}

Using (5),(6),(7)(5), (6), (7) to simplify (4)(4) we get

λ22fλ+f2+μ22gμ+g2=r2det(A) \lambda^{2} - 2f\lambda + f^{2} + \mu^{2} - 2g\mu + g^{2} = r^{2} - \det(A)

which is

(λf)2+(μg)2=ρ2\begin{equation} (\lambda - f)^{2} + (\mu - g)^{2} = \rho^{2} \end{equation}

where ρ2=r2det(A)\rho^{2} = r^{2} - \det(A). (8)(8) gives the equation for the eigencircle of AA.

Deriving matrix properties from Eigencircles

Okay, now consider (3)(3) again. We know that the determinant of any matrix with a zero row or a zero column is zero. Considering the values of λ,μ\lambda, \mu that would produce zero rows and columns, we get the following points: (a,b),(d,b),(a,c),(d,c)(a,b), (d,b), (a,-c), (d,-c). These points must be on the eigencircle since they satisfy (3)(3). Moreover, since the center of our eigencircle is defined to be the midpoint of (a,d)(a,d) and (b,c)(b,-c)

(f,g)=(a+d2,bc2)(f,g) = (\frac{a+d}{2}, \frac{b-c}{2})

these points define a rectangle.

Defining the points F=(a,b),H=(a,c),E=(d,c),G=(d,b)F = (a,b), H = (a,-c), E = (d,-c), G = (d,b), we get the following graph.

eigencircle
Figure 2: An eigencircle (assuming a < d and b < -c)

Before moving forward, you may notice the axes on Figure 2 corresponding to x,yx, y - these diagrams are from the original article and I was too lazy to refactor them. For the purposes of this article, x=λx = \lambda and y=μy = \mu.

This visualization has some very interesting properties. Firstly, notice that if (λ,0)(\lambda, 0) lies on the circle, then λ\lambda is an eigenvalue of AA. Let L=(λ,0)L=(\lambda, 0), then the vector LE\overrightarrow{LE} is the corresponding eigenvector for λ\lambda.

eigencircle

Figure 3: Determining eigenvectors from eigencircles

To see why, note that the vector LE=(dc)(λ0)=(dλc)\overrightarrow{LE} = \begin{pmatrix} d \\ -c \end{pmatrix} - \begin{pmatrix} \lambda \\ 0 \end{pmatrix} = \begin{pmatrix} d - \lambda \\ -c \end{pmatrix}. If this is an eigenvector corresponding to eigenvalue λ\lambda, then it must be in the nullspace of the matrix [aλbcdλ]\begin{bmatrix} a - \lambda & b \\ c & d - \lambda \end{bmatrix}.

Let’s see if this is the case:

[aλbcdλ](dλc)=((aλ)(dλ)bc0)\begin{bmatrix} a - \lambda & b \\ c & d - \lambda \end{bmatrix} \begin{pmatrix} d - \lambda \\ -c \end{pmatrix} = \begin{pmatrix} (a-\lambda)(d-\lambda)-bc \\ 0 \end{pmatrix}

But we know that (aλ)(dλ)bc=0(a-\lambda)(d-\lambda)-bc = 0 from (3)(3) (since (λ,0)(\lambda, 0) lies on the eigencircle). Therefore, LE\overrightarrow{LE} is indeed an eigenvector correspoinding to λ\lambda.

We can also use the eigencircle to prove that the product of eigenvalues for a 2x2 matrix AA is equal to the value of its determinant. We use the “power of a point” theorem to show this. The power of any point PP given by Π(P)\Pi(P) with respect to some circle cc with center OO is defined as

Π(P)=PO2r2\begin{equation} \Pi(P) = |PO|^2 - r^2 \end{equation}

Geometrically, the power of a point states that “[if some] line through any point PP meets a circle in points QQ and RR, the product of the lengths, PQPQ and PRPR, is the same for any direction of the line. The lengths are signed, so the product is positive (negative) when PP is outside (inside) the circle.”

eigencircle

Figure 4: Power of a point theorem

Consider Figure 4. What the power of a point theorem says is that the (signed) product of the lengths PXPX=PYPY=Π(P)PX \cdot PX' = PY \cdot PY' = \Pi(P). For more on the power of a point, refer to this link.

Consider (9)(9). For the eigencircle, we can rewrite this as (Pxf)2+(Pyg)2ρ2(P_x - f)^2 + (P_y - g)^2 - \rho^2. Notice that this expression can be expressed as the LHS of (3)(3) by construction. So we get that the power of a point with respect to a eigencircle is given by

Π(P)=(aPx)(dPx)(bPy)(c+Py)\begin{equation} \Pi(P) = (a-P_x)(d-P_x)-(b-P_y)(c+P_y) \end{equation}

If we choose P=O=(0,0)P = O = (0,0), we get that Π(O)=det(A)\Pi(O) = \text{det}(A)! From Figure 3, we know that the (signed) lengths of the line segments OL1OL_1 and OL2OL_2 specify the eigenvalues. Well, from the power of a point theorem, we know that OL1OL2=Π(O)=det(A)OL_1 \cdot OL_2 = \Pi(O) = \text{det}(A). Personally, I found this fascinating.

Finally, I am going to describe how the eigencircle relates to complex eigenvalues. Consider Figure 5a.

eigencircle

Figure 5: Eigencircles and complex eigenvalues

From the power of a point theorem, we know that the power of YY is given by

Π(Y)=YL1YL2=YMYN\Pi(Y) = YL_1 \cdot YL_2 = YM \cdot YN

As YL1=YL2YL_1 = -YL_2, we have that YL12=YMYN>0YL_1^2 = YM \cdot YN > 0 (remember that we consider signed lengths). Knowing this, we can express the eigenvalues (L1,L2L_1, L_2) in the following form

λ=OY±YMYN\begin{equation} \lambda = OY \pm \sqrt{-YM \cdot YN} \end{equation}

Now this is also valid for cases where the x-axis does not intersect the eigencircles (we will see why in a minute). Consider Figure 5b. We see that the quantity YMYNYM \cdot YN is now positive, there the value inside the square in (11)(11) is negative, giving us a complex value. If YVYV is a tangent to the circle at VV, then, by the power of a point theorem, we have that YMYN=YV2YM \cdot YN = YV^2. Hence, we get the complex eigenvalues

λ=OY±iYV\lambda = OY \pm iYV

Okay, so why is (11)(11) true when the x-axis does not intersect the eigencircle? Let’s figure out the coordinates Y,M,NY,M,N. Well, we can see that Y=(f,0),M=(f,gρ),N=(f,g+ρ)Y = (f, 0), M = (f, g-\rho), N = (f, g+\rho) (recall definitions 5,6,75, 6, 7). Plugging this into (11)(11) we get

λ=f±(g2ρ2)\lambda = f \pm \sqrt{-(g^2-\rho^2)}

or equivalently,

λ=f±ρ2g2\begin{equation} \lambda = f \pm \sqrt{\rho^2 - g^2} \end{equation}

Notice that if the x-axis does not intersect the eigencircle, then we have that ρ2<g2\rho^2 < g^2 and therefore ρ2g2<0\rho^2 - g^2 < 0, just as we had derived earlier. For this case, we need to realize that ρ2g2\sqrt{\rho^2 - g^2} is undefined and we need to factor out the 1\sqrt{-1}. More on this below.

For λ\lambda to be an eigenvalue, we know that equation (3)(3) must be satisfied for (λ,0)(\lambda, 0). Notice that equation (3)(3) can be rewritten as

(λf)2+(μg)2ρ2=0\begin{equation} (\lambda - f)^2 + (\mu - g)^2 - \rho^2 = 0 \end{equation}

This is just (8)(8) with the ρ2\rho^2 moved to the LHS. Plugging in (λ,0)(\lambda, 0), we get

(f±ig2ρ2f)2+g2ρ2=0(±ig2ρ2)2+g2ρ2=0(g2ρ2)+g2ρ2=0ρ2g2+g2ρ2=00=0\begin{align*} \biggr(f \pm i\sqrt{g^2 - \rho^2} - f\biggr)^2 + g^2 - \rho^2 &= 0 \\ \biggr(\pm i\sqrt{g^2-\rho^2}\biggr)^2 + g^2 - \rho^2 &= 0 \\ -(g^2 - \rho^2) + g^2 - \rho^2 &= 0 \\ \rho^2 - g^2 + g^2 - \rho^2 &= 0 \\ 0 &= 0 \\ \end{align*}

as required. Therefore, you can determine all (both real and complex) eigenvalues using the eigencircle.

Visualizing the eigenvectors corresponding to complex eigenvalues would require an additional axis to account for the imaginary value. Let λ=f±ih\lambda = f \pm ih where h=YVh = YV. Consider Figure 6.

eigencircle

Figure 6: Geometric representation of complex eigenvectors

LE\overrightarrow{LE}

and

KE\overrightarrow{KE}

represent the complex eigenvectors. Notice

LE=OEOL=(dc0)(f0h)\overrightarrow{LE} = \overrightarrow{OE} - \overrightarrow{OL} = \begin{pmatrix} d \\ -c \\ 0 \end{pmatrix} - \begin{pmatrix} f \\ 0 \\ h \end{pmatrix}

represents the complex eigenvector (d(f+ih)c)=(dfihc)\begin{pmatrix} d - (f + ih) \\ -c \end{pmatrix} = \begin{pmatrix} d - f - ih \\ -c \end{pmatrix}.

I am going to stop there. It was fun reading about eigencircles. I believe the way in which the properties of eigenvectors and eigenvalues expressed themselves through geometry was beautiful. Refer to [1] and [2] for the original articles on eigencircles.



References

  1. Englefield, M. J., & Farr, G. E. (2006). Eigencircles of 2 x 2 Matrices. Mathematics Magazine Vol. 79 Oct.,2006, 281-289.
  2. Englefield, M. J., & Farr, G. E. (2010). Eigencircles and associated surfaces. The Mathematical Gazette Vol.94 No. 531 (November 2010), 438-449.