Chapter 8 — Eigenvalues and Eigenvectors

Chapter 8 — Eigenvalues and Eigenvectors

Up to this point, we’ve explored the tools that help us solve linear systems, decompose matrices, and build numerically stable foundations for real-world computation. But now we cross a threshold into one of the most powerful, far-reaching, and conceptually beautiful ideas in all of applied mathematics:

Eigenvalues and eigenvectors.

Few concepts appear in as many places as these do. From principal component analysis to mechanical vibrations, from PageRank to quantum mechanics, from stability analysis in control systems to optimization in machine learning—eigenvalues quietly determine how systems behave, evolve, and respond.

And yet, in practice, most people learn eigenvalues in the most abstract possible way:

  • A x = λ x
  • Find det(A − λI) = 0
  • Solve the characteristic polynomial

This feels like mathematics done purely on paper—a sequence of symbolic steps without connection to what computers actually do. But the truth is quite different. In numerical computation, eigenvalues are not found by solving polynomials. They are found by iterative algorithms, matrix transformations, and clever exploitation of structure. Every modern scientific computing stack (LAPACK, NumPy, Julia, MATLAB, ML frameworks) uses a handful of deeply optimized routines whose origins lie in a mixture of geometry, algebra, and numerical stability considerations.

This chapter is about understanding those ideas—not from a symbolic algebra perspective, but from the viewpoint of a computational engineer trying to build systems that behave predictably under real constraints.


Why do eigenvalues matter so much?

There’s a subtle but profound reason eigenvalues show up everywhere:

Eigenvalues describe what a matrix does when it acts repeatedly.

If you multiply a matrix by a vector once, you get a new vector. If you multiply it again, you get the next step in a dynamical system. Repeat this many times, and the dominant eigenvalue—and its associated eigenvector—tell you where the system is going.

This insight underlies:

  • PCA and low-rank approximation
  • Graph algorithms like PageRank
  • Stability in differential equations
  • Markov chains
  • Kernel methods
  • Deep learning initialization and training behavior
  • Optimization dynamics

The spectral properties of matrices are not just theoretical. They govern the behavior of real algorithms.


The computational challenge

So how do we compute eigenvalues in practice? Not by solving characteristic polynomials—those explode numerically even for moderately sized matrices. Instead, numerical linear algebra relies on:

  • Power iteration to extract dominant eigenvalues.
  • Inverse iteration and Rayleigh quotient iteration to find precise eigenpairs.
  • The QR algorithm, one of the great achievements of numerical mathematics, used for full spectrum computation.
  • Specialized structures and transformations (Hessenberg form, orthogonal similarity) that make computation stable and efficient.

In other words:

Eigenvalue computation is not algebra. It is engineering.

And to understand that engineering is to unlock an entirely new understanding of how algorithms behave.


Where eigenvalues show up in modern computation

This chapter also looks outward—to the many fields where spectral methods are essential:

  • PCA: find directions of maximum variance.
  • SVD: which generalizes eigenvalue concepts for rectangular matrices.
  • Graph Laplacians: clustering, diffusion processes, community detection.
  • Kernel methods: eigendecomposition of Gram matrices.
  • Gaussian processes: covariance matrices, spectral smoothness.
  • Neural networks: studying training dynamics and mode dominance.

Many “advanced” ideas in ML and data analysis are, at their core, just clever applications of eigenvalues and eigenvectors. By understanding the mechanics here, we gain leverage across countless areas.


What this chapter covers

This chapter is structured around four major ideas:

  1. Power method and inverse iteration: how repeated multiplication extracts eigenpairs.
  2. Rayleigh quotient: a remarkably simple tool that supercharges convergence.
  3. The QR algorithm (high-level intuition): the backbone of modern eigenvalue computation.
  4. PCA and spectral methods: practical applications where eigenvalues make systems interpretable.

Each topic builds toward an understanding of how spectral information is extracted, interpreted, and used in real-world systems. The emphasis is always on:

  • intuition
  • geometric understanding
  • numerical stability
  • practical computational behavior

We will not shy away from mathematical clarity, but we will keep the discussion grounded in real computation, just as in earlier chapters.


A natural step forward

We’ve already seen, in earlier chapters, how QR decomposition stabilizes problems, exposes structure, and provides numerical safeguards. Now we extend that intuition into a new domain—one where QR is not just useful, but central.

Before diving into the QR algorithm itself, we begin with the most fundamental question in spectral analysis: How can repeated multiplication reveal the hidden structure of a matrix?

Next: Power Method and Inverse Iteration

2025-10-06

Shohei Shimoda

I organized and output what I have learned and know here.