Re: algebra presentations
Well, loosely speaking, it means we are going to take every single polynomial that has $X^2 + 1$ as a factor, and set it to 0.
Before we go there, however, let's look at a less complicated example:
Algebraically, the integer 0 has 2 unique properties other integers do not have:
0 + 0 = 0
k * 0 = 0
Now...suppose we wanted to find some subset J of the integers that also had these properties:
$J + J = J$
$\Bbb ZJ = J$.
OK, well, first of all, how do we even "add" sets?!?
So we have to give some MEANING to $J + J$. Here is what we will do:
$J + J = \{k + m: k,m \in J\}$
That seems straight-forward enough, yes?
So when we say: $J + J = J$, we mean two things:
1) $J$ is closed under addition,
2) every element in $J$ is the sum of two other elements in $J$.
Similarly, by $\Bbb ZJ$ we mean the set $\{ab: a\in \Bbb Z, b\in J\}$.
One possibility that immediately springs to mind is the set of all multiples of $k$, for any given integer $k$. Clearly it satisfies (1), since:
$km + kn = k(m+n)$
and since $0 = k0$, we can write $km = km + k0$, both of which are multiples of $k$.
Clearly this set "absorbs" the integers much like 0 does, if we multiply any multiple of $k$ by any integer, we still wind up with a multiple of $k$:
$n(km) = k(nm)$.
so this set does the second thing we want, as well.
So such a set could possibly be a "zero-like object" in some "integer-like" system. What would treating every multiple of $k$ as "some kind of 0-thing" do to our integer system?
Well, counting would go like this:
0,1,2,...,k-1,k (back to 0)
k+1 (same as 1),k+2 (same as 2),...,2k-1 (same as k-1), 2k (back to 0)
2k+1 (same as 1), 2k+2 (same as 2)...etc.
In other words, the "line" of integers, has turned into a "cycle" which repeats every $k$ "clicks".
Formally, what we do is this:
Suppose $R$ is a commutative ring, with a subset $J$ such that:
1) $(J,+)$ is an additive subgroup of $(R,+)$ (such a subgroup satisfies $J + J = J$, by the closure property of group operations)
2) If $a \in R$ and $x \in J$, then $ax \in J$ (this means that $RJ = J$, the "absorbing property").
We call such a set $J$ an IDEAL of $R$ (it's a strange name, but let's just go with it).
Here is where things get interesting:
We can define an equivalence relation on $R$ by:
$a \sim b \iff a - b \in J$.
Let's verify this is indeed an equivalence relation.
Is $a \sim a$?
Well, $a - a = 0$, and since $J$ is an additive subgroup, it surely contains the additive identity, so yes, $\sim$ is reflexive.
Suppose $a \sim b$. Is it always true that $b \sim a$?
$a \sim b$ means that $a - b \in J$. Since $J$ is an additive group, this means $-(a - b) = b - a \in J$ as well, so yes!, this relation is symmetric, too.
Finally, suppose $a \sim b$ and $b \sim c$. This means $a - b \in J$, and $b - c \in J$. Since $J$ is closed under addition, we have:
$(a - b) + (b - c) = a - c \in J$, so $\sim$ is transitive, as well.
So we really do have a bona-fide equivalence relation. That said, what does the equivalence class of a given $a \in R, [a]$ look like?
Well, suppose $x \in J$ is any element of $J$. Then $a + x - a = x \in J$, so:
$a \sim a+x$. So clearly we have:
$a + J = \{a + x: x \in J\} \subseteq [a]$.
On the other hand, suppose $b$ is any element of $[a]$. Since:
$b = a + b - a$, and $b \sim a, b - a \in J$ so $b \in a + J$.
This means that the equivalence class of $a$ IS the set $a + J$.
Now, one question is...can we turn these equivalence classes into a ring themselves? To do that, we need to come up with some way of adding them, and multiplying them, and then we need to check that the ring axioms hold.
Well, it may or may not seem obvious, but one way is to define:
$[a] + = [a + b]$, or equivalently, $(a + J) + (b + J) = (a+b) + J$.
$[a] = [ab]$, or: $(a + J)(b + J) = (ab) + J$.
There is one small problem, however: $[a]$ may have a LOT of elements in it, and so we could have:
$[a] = [a']$ (and likewise for $$)
So if we're going to use $a$ and $b$ to define the sum and product of the equivalence classes, we ought to make sure that if:
$[a] = [a'], = [b']$
that $[a+b] = [a'+b']$ and $[ab] = [a'b']$, or our definition isn't going to be CONSISTENT.
Now $[a] = [a']$ means that $a \sim a'$, which is to say $a - a' \in J$. Similarly $ = [b']$ means that $b - b' \in J$.
So $a + b - (a' + b') = (a - a') + (b - b') \in J$, since the sum of two elements of $J$ is always again in $J$ (see how that $J + J = J$ rule is working for us, now?). This means that $[a + b] = [a' + b']$, whenever we have $[a] = [a'], = [b']$, so our sum doesn't depend on the particular $a,b$ we use, just the equivalence class. That's reassuring.
Now suppose $a \sim a', b \sim b'$. Then:
$ab - a'b' = ab - ab' + ab' - a'b' = a(b - b') + (a - a')b'$
Now $b - b' \in J$ (since $b \sim b'$), and since anything multiplied by something in $J$ is again in $J$, we have:
$a(b - b') \in J$.
Similarly, $(a - a')b' = b'(a - a') \in J$, since $a - a' \in J$.
So $ab - a'b'$ is the sum of two elements of $J$, so must ALSO be in $J$. This means we can trust that our product of equivalence classes also makes sense, no matter which "representatives" we use to calculate it.
Now...this is all a bit abstract, so let's go back to our integer example:
$J = k\Bbb Z = \{kn: n \in \Bbb Z\}$.
We see that our equivalence classes are:
$k\Bbb Z = \{0,k,-k,2k,-2k,3k,-3k,\dots\}$
$1 + k\Bbb Z = \{1,k+1,-k+1,2k+1,-2k+1,3k+1,-3k+1,\dots\}$
...
$k-1 + k\Bbb Z = \{k-1,2k-1,-1,3k-1,-k-1,4k-1,-2k-1,\dots\}$
and that our equivalence class $[m]$ is just the integer $m$ modulo $k$.
*********
Now polynomials are sort of like "grown-up integers", the algebra for them is VERY similar. Instead of using an ideal generated by an integer $k$ like we did for the integers, we're going to use an ideal generated by the polynomial $X^2 + 1$.
Now there's a LOT of multiples of $X^2 + 1$, like:
$X^3 + X$
$X^3 + 2X^2 + X + 2 = (X + 2)(X^2 + 1)$ etc.
We're going to take ALL these multiples and call them...(I hope you can guess this) $J$.
So given a polynomial...any old polynomial, instead of considering the polynomial $p(X)$, we're going to look at the equivalence class:
$[p(X)] = p(X) + J$.
Well, we're going to need to write some of these equivalence classes down, in order to compute with them, and we don't want to have to use really BIG polynomials (that would be HARD). So what can we do? Hmm...
Suppose, just suppose, we divided $p(X)$ by $X^2 + 1$, and we got something like:
$p(X) = q(X)(x^2 + 1) + r(X)$.
Since $r(X)$ is a remainder, we could keep going until we got a remainder with degree less than the degree of $X^2 + 1$, that is:
$r(X) = aX + b$.
Now:
$p(X) - r(X) = q(X)(X^2 + 1)$, and guess what? $q(X)(X^2 + 1) \in J$.
This means:
$[p(X)] = [r(X)]$.
So we catch a break, here...instead of working with ARBITRARY polynomials, we can just use their REMAINDERS upon division by $X^2 + 1$, and these remainders are going to be SMALL polynomials, which makes everything a LOT easier.
Now suppose:
$[p_1(X)] = [r_1(X)] = [a + bX]$ and:
$[p_2(X)] = [r_2(X)] = [c + dX]$.
We know that:
$[p_1(X)] + [p_2(X)] = [r_1(X)] + [r_2(X)] = [a + bX] + [c + dX]$
$= [a + bX + c + dX] = [(a + c) + (b + d)X]$
so adding in this system of "equivalent" polynomials is pretty easy. What about multiplying?
We'll try to use our "simple" representatives to make the math easier:
$[p_1(X)][p_2(X)] = [r_1(X)][r_2(X)] = [a + bX][c + dX]$
$= [(a + bX)(c + dX)] = [ac + (ad)X + (bc)X + (bd)X^2]$
$= [ac + (ad + bc)X + bdX^2]$.
Huh.
Our product has an $X^2$ term, which is kinda gumming up the works. We should probably find out what the equivalence class of $X^2$ is, since it's of degree higher than 1. Ok.
Now, $X^2 = X^2 + 0 = X^2 + 1 - 1 = -1 + (X^2 + 1)$.
This tells us: $X^2 + J = -1 + J$, that is: $[X^2] = [-1]$. This is helpful.
So:
$[p_1(X)][p_2(X)] = [ac + (ad + bc)X + (bd)X^2]$
$= [ac] + [(ad + bc)X] + [(bd)X^2] = [ac] + [(ad + bc)X] + [bd][X^2]$
$= [ac] + [(ad + bc)X] + [bd][-1] = [ac] + [(ad + bc)X] + [-bd]$
$ = [ac - bd] + [(ad + bc)X] = [(ac - bd) + (ad + bc)X]$.
There! now we have the "reduced" form of a product.
So how is this related to $\Bbb C$?
What we do is map the equivalence class:
$[p(X)] = [a + bX] \mapsto a+bi \in \Bbb C$.
This is essentially "setting all multiples of $X^2 + 1$ equal to 0" since:
$J = 0 + J = [X^2 + 1] = [0] = [0 + 0X] \mapsto 0 + 0i = 0$.
($X^2 + 1$ has a remainder of 0 when divided by itself, right?)
What happens to $[X]$ under this map?
It gets sent to the complex number $i$, and $i^2 + 1 = 0$
(in other words, the equivalence class of $X$ relative to the multiples of $X^2 + 1$ actually is a SOLUTION to the polynomial:
$X^2 + 1$
since:
$[X]^2 + [1] = [X^2 + 1] = [0]$).
Geometrically (although this is not quite accurate), you can imagine the polynomial space $\Bbb R[X]$ as a big blob, that we slice into "spaghetti strings" of translates of the multiples of $X^2 + 1$. The strings containing just $[X]$ terms map to points on the $y$-axis, the strings containing just $[a]$ terms (the polynomials equivalent to constant polynomials) map to points on the $x$-axis (the reals...after all, constant term polynomials act just like the reals), so the complex numbers are like a cross-section through this mass of spaghetti strings (which are pretty thin, by the way), so that we get exactly one complex number per string. The string containing $X^2 + 1$ goes through the origin, of course.