Home> urticator.net Search About This Site > Domains Glue Stories Computers Driving Games Humor Law > Math Numbers Science > Continued Fractions Game Theory (Section) Group Theory Probability Miscellaneous Approximation Quality Two Examples The Markov Equation Equivalence The Markov Constant The Pattern Group Elements and Matrices The Random Thoughts Symmetries Chains and Denominators > The Discriminant Properties of Rotations Reflection Symmetry A Nice Correspondence Epilogue
|
One More DiscriminantUnfortunately, we're not done yet. I didn't point it out at the time, but the discriminant appeared in one more place. Remember how in The Pattern we had an exact equation for chains 5a and 5b?
In that context, u and v varied, w = 5 was constant, and the solutions were unnormalized. I can't explain the next step better than I did before.
See how the solutions loop around through the tree? At the center of the loop (n = 0) is a solution that when normalized has u = 5, … That solution has discriminant 9u2 − 4 = 221, and that gives us one more connection to establish. Fortunately, we already have most of the tools we'll need, but I do need to say a bit more about the theory of recurrence relations. So, suppose we have a recurrence relation with some characteristic polynomial. For each root x of the polynomial, the sequence xn is a solution of the recurrence relation, and, ignoring some complications about multiple roots, every solution of the recurrence relation is a linear combination of those basic solutions. The weights can be computed from the roots and the initial conditions. Here's a fact that I've used in other places but won't need in this essay. The basic solutions xn grow exponentially because the variable is n, not x, so in a linear combination, the solution with the largest root rapidly dominates the rest. That's why the largest root is the most important.
Anyway, now we can parse the exact equation: the roots are (15 ± root 221)/2 and the weights are (1/2) ± (11/442) root 221. We can also answer some questions. Why do the weights contain the discriminant? Because they're computed from the roots and the initial conditions, and the roots contain the discriminant. And why do the roots contain the discriminant? Because they're the roots of the quadratic equation
x2 − 3ux + 1 = 0, and the discriminant b2 − 4ac of that quadratic equation is … 9u2 − 4. That seems like it ought to mean that we're done, but no. What it means is that the appearance of the discriminant in the exact equation for chains 5a and 5b wasn't some kind of coincidence. For every value of u, the quadratic equation above and the equation for the value of expansion(u) have the same discriminant, even though they're different equations. The question now is why those two equations have the same discriminant. For comparison, here's the equation for the value of expansion(5) from Group Elements and Matrices.
5z2 − 9z − 7 = 0 For a better comparison, let's generalize. The characteristic polynomial
x2 − 3ux + 1 originally came from the Markov equation, which as far as I know doesn't generalize, but later (in Chains and Denominators, specifically The Recurrence Relation) we saw that it's also the characteristic polynomial of the recurrence relation for the subsequence s−1 of the denominators of expansion(u), and that does generalize to arbitrary expansions.
x2 − (tr G) x + (det G) Then we can write that out in terms of matrix entries, turn it into a quadratic equation, …
x2 − (a + d)x + (ad − bc) = 0 … and compare it to the equation for the value of an arbitrary expansion.
cz2 − (a − d)z − b = 0 So, yes, the equations are different. For reference, here's the generalized version of the discriminant (again).
D = (tr G)2 − 4 det G Next I want to introduce a new element and prove that the roots of the characteristic polynomial are equal to the eigenvalues of the matrix G. In fact, I want to prove it twice. The first proof is simple but kind of disappointing. All we need to know is that the determinant of a matrix is the product of the eigenvalues, and the trace is the sum.
A quick note about names: for any matrix, the polynomial with the eigenvalues as roots is called the minimal polynomial and the polynomial with the eigenvalues as roots with correct multiplicity is called the characteristic polynomial. So, here, the characteristic polynomial of the recurrence relation is also both the minimal polynomial and the characteristic polynomial of the matrix. The second proof is kind of long, so I'll just sketch it out. Consider the subsequence s−1 of the denominators of the expansion. On the one hand, we know that the growth of the subsequence is controlled by the roots of the characteristic polynomial. On the other hand, it turns out that the growth of the subsequence is also controlled by the eigenvalues of G, so the two must be identical. How does it turn out that way? Well, remember equations (1) and (2) in The Recurrence Relation? Instead of using them to do algebra, we can just regard them as a single matrix equation. That equation says that if we multiply the row vector (r−1 r−2) by G (on the right, of course), we get the row vector (rL−1 rL−2). In other words, multiplying by G advances the solution vector by L. So, if we multiply by G repeatedly, we can compute the whole subsequence s−1, and if we decompose the solution vector into eigenvectors, we can see that the growth is controlled by the eigenvalues. That matrix equation seems strange, so let's take a minute to see how it fits into the scheme of things. So far we've seen a lot of matrices, but really only one nontrivial matrix equation, the result at the end of Group Elements and Matrices. In the separate essay A Nice Correspondence I generalized that result into what I'm now calling the first property of the prefix function P.
P(A & B) = P(A)P(B) Let's look at two other special cases of that property.
All that's left now is to show that the eigenvalues of the matrix G and the value of the expansion have the same discriminant, but that's easy. Suppose we have an eigenvector that's a column vector. Because c ≠ 0, the second component can't be zero, so we can rescale the eigenvector to make the second component 1. If we call the first component z, then the eigenvector equation Mv = λv looks like this.
On the one hand, we can divide the first equation by the second and see that z is a fixed point of G(z). On the other hand, we can just look at the second equation and see that z and λ are related by a linear equation with integer coefficients and therefore have the same discriminant. In particular, the eigenvalue λ+ is related to the fixed point z+, which is the value of the expansion. Q.E.D. As confirmation, here are the fixed points and eigenvalues from Group Elements and Matrices.
That's the end of the random thoughts and the end of giant thread, but be sure to check the table of contents at left for other closely related essays, for example Properties of Rotations.
|
See Also@ January (2024) |