453 Notes
453 Notes
Uli Walther
c 2021
Version of Spring 2021
Contents
Basic notions 7
0.1. How to use these notes 7
0.2. Set lingo 7
0.3. Size of sets 8
0.4. Finite vs infinite 9
0.5. Inclusion/Exclusion 10
1. Making quotients 57
2. The isomorphism theorem 60
Chapter VII. Week 7: Finitely generated Abelian groups 63
1. Row reduced echelon form over the integers 63
2. Generating groups 66
Expected progress:
• Week 1: Archimedes, Factorization
• Week 2: Symmetries, Groups, Subgroups, Order, Aut(G)
• Week 3: Z/nZ, Products, U (n)
• Week 5: Cosets, Morphisms
• Week 4: Symmetric and Free Group
• Week 6: Normal Subgroups, Quotients, Automorphism Theorem
• Week 7: Finitely Generated Abelian Groups
• Week 8: Review, Group Actions
• Week 9: Reading day, Intro to Rings
• Week 10: Midterm, Ideals, Morphisms
• Week 11: Euclidean Algorithm, PID, UFD
• Week 12: Fields, Eisenstein, Extensions,
• Week 13: Degrees and Splitting Fields
• Week 14: Minimal Polynomials, Finite Fields
• Week 15: Galois Outlook, Review
Basic notions
0.1. How to use these notes. These notes contain all I say in class, plus
on occasion a lot more. If there are exercises in this text, you may do them but
there is no credit, and you need not turn them in. All exercises that are due are
specifically listed on gradescope.
This initial chapter is here so we have acommon understanding of the basic
symbols and words. This should be known from MA375 (at least if I teach it).
Chapter 1 is still more than we did in week 1, but almost all of it should be
familiar, and the rest (the open problems on primes) is for your entertainment.
Future chapters correspond to actual weeks of classes and are much less verbose
than the Basic Notions.
The chapter “Stuff for later” can be ignored. It will be worked in when the
time comes.
Remark .1. There are typos in these notes. If you find some, please inform
me.
0.2. Set lingo. The mathematical word set denotes what in colloquial life
would be called a collection of things. The “things” are in mathematics referred to
as elements of the set. If S is a set and s is an element of S then one writes s ∈ S.
A sub-collection S 0 of elements of S is a subset and one writes S 0 ⊆ S, allowing
for the possibility of S 0 being all of S, or to have no element. There is, strangely,
a set that contains nothing. It’s called the empty set (denoted ∅) and, despite
its humble content, one of the most important sets. One uses the notation S =
{s1 , . . . , sn , . . .} to indicate that S consists of exactly the elements si . (In many
cases, one must allow the index set to be different from N. In other words, not all
sets can be “numbered”, a fact we explore a bit below).
A function φ : A → B from the set A to the set B is an assignment (think:
a black box) that turns elements of A into elements of B. The crucial conditions
are: the assignment works for every single input a ∈ A (so the black box does not
choke on any input from A) and for each input there is exactly one output specified
(no more, no less). Graphically, functions are often depicted by the help of arrows
(starting at the various elements of A and ending at the value φ(a) for each input
a). (For an example, suppose φ : Z → Z is the process of doubling. Then one could
write φ(3) = 6, or 3 7→ 6). The set A is usually referred to as source, the set B as
target.
Definition .2. The function φ : A → B is
(1) surjective (“onto”) if every element b appears as output of φ;
(2) injective (“into”) if the equality of outputs φ(a1 ) = φ(a2 ) occurs exactly
when the inputs a1 and a2 were equal;
(3) bijective if it is injective and surjective.
7
8 BASIC NOTIONS
0.3. Size of sets. We wish to attach to each set S a size denoted |S|. In order
to make sense of this, we need to compare sets by size.
Definition .3. We write |S| ≤ |S 0 | if there is an injective map φ : S ,→ S 0 .
Do not confuse the symbols ≤ and ⊆. The following examples illustrate the
nature of the relation ≤.
Example .4.
|N| ≤ |Z| since each natural number is an integer.
Exercise .5. Show that |Z| ≤ |N|.
Example .6. • |Z| ≤ |Q| since each integer is a rational number.
• |Q| ≤ |R| since each rational number is also real.
• Somewhat shockingly, |Q| ≤ |Z|. To see this, it will be sufficient to prove
that there is a way of labeling the rational numbers with integer labels. (One can
then make an injective map that sends each rational to its label). How does one
label? Imagine sorting the rational positive numbers into a two-way infinite table,
as follows:
q
.
p .. 1 2 3 4 ···
1 1/1 1/2 1/3 1/4 · · ·
2 2/1 2/2 2/3 2/4 · · ·
3 3/1 3/2 3/3 3/4 · · ·
4 4/1 4/2 4/3 4/4 · · ·
.. .. .. .. .. ..
. . . . . .
Clearly all positive rationals appear (multiple times) in the table. Now suppose
you are moving through the table “on diagonals” where p + q is constant: start at
1/1, the only square on its diagonal (where p + q = 2). Next go on the diagonal
p + q = 3, starting on the square with 1/2 and then moving down and to the left.
Next walk along the diagonal p + q = 4 starting on 1/3 and moving down and left.
It is clear that this process allows you to label each field: 1/1 is number 1, 1/2 is
number 2, 2/1 is number 3, and so on. So, the set of all squares is in bijection with
the set N. Since all positive rationals are sorted into the various fields, it follows
that |Q| ≤ |{all squares}| ≤ |N|. A similar idea can be used on negative numbers,
and this shows that |Q| ≤ |Z|.
• In contrast, the statement |R| ≤ |Q| is false. The idea is due Cantor, and
goes like this. If you believe that you can inject R into Q then you can also inject
R into Z because |Q| ≤ |Z|. Since |Z| ≤ |N|, this also implies that you can inject R
into N. To inject R into N means to label the real numbers by using only natural
(non-repeated) indices. In particular, this can be done to the reals between 0 and
1.
BASIC NOTIONS 9
by |2S | = 2|S| . If S is infinite, such equation makes no sense. But in any event, 2S
is strictly larger than S in the sense that there is no injection 2S ,→ S. The idea of
the proof is the same as the Cantor diagonal trick for S = N we saw above.
Exercise .8. If the set S is finite, prove that |2S | = 2|S| . (Hint: an element
S
of 2 is a subset of S. What question do you need to answer for each element of S
when you form a subset of S? How many possible answers can you get?)
Exercise .9. Let S be a finite set of size n. Determine (in terms of n) the
number of pairs of sets (A, B) where both A and B are subsets of S, and where no
element of S is both in A and B. Prove the formula you find.
So, for example, if S has one element called s, then the options for (A, B) are:
(∅, ∅), (∅, {s}) and ({s}, ∅).
0.5. Inclusion/Exclusion.
Notation .11. Given two sets A and B, their union is the set A ∪ B that
contains any element in A, any element in B, and no other. On the other hand,
the intersection A ∩ B is the set that contains exactly those elements of A that are
also in B, and no other.
Tk
For a list of sets A1 , . . . , Ak their common intersection is denoted i=1 Ai and
Sk
their union i=1 Ai .
Suppose A and B are two finite sets; we want to know the size of their union
A ∪ B. A first order approximation would be |A| + |B|, but this is likely to be off
because A and B might have overlap and elements in the overlap A ∩ B would be
counted twice, once in A and once in B. So, we must correct the count by removing
one copy of each element in the overlap:
How about three sets? In that case, there are three intersections: A ∩ B, B ∩ C
and A ∩ C, whose sizes should presumably all be removed from |A| + |B| + |C|.
This is the right idea but doesn’t quite capture it. For example, if A = {1, 2, 3},
B = {3, 4} and C = {2, 3, 5} then |A| + |B| + |C| − |A ∩ B| − |A ∩ C| − |B ∩ C| is
3 + 2 + 3 − 1 − 1 − 2 = 4 while the union is the set {1, 2, 3, 4, 5}. To understand
what happened, look at each element separately. The expression above counts each
of 1, 2, 4, 5 a total of once. But the element 3 is counted three times, and then
removed three times. So, the count is off by one. Inspection shows that this error
will always happen if the intersection A ∩ B ∩ C is not empty, and the count will
be off by as many elements as this intersection contains. So, we conclude:
Week 1: Introduction
1. Induction
Suppose you are faced with the task of proving that, for all natural numbers
n, the sum 1 + 2 + . . . + n equals n(n + 1)/2. A few tests show that the formula is
probably right, but no matter how many checks you do, there are infinitely many
others yet to run. It seems like a hopeless proposition. Mathematical induction is
a tool that allows you to complete precisely this sort of job.
1.1. Setup. Suppose that, for each n ∈ N, there is given a statement P (n)
that involves the symbol n more or less explicitly. (For example, P (n) could be the
statement “the sum 1 + 2 + . . . + n equals n(n + 1)/2” from above).
The task at hand is to prove that all statements P (0), P (1), . . . are true.
1.2. The idea. Imagine you are standing in front of a ladder that starts at
your feet (level 0) and goes up indefinitely. Your job is to convince your friend that
you are capable of climbing up to any step of the ladder. How might you do that?
13
14 I. WEEK 1: INTRODUCTION
One approach is to check that you can indeed make it to the lowest rung of the
ladder (the “base case”) and then to exhibit a kind of cranking mechanism that
allows for any position on the ladder (no matter which exact one), say rung n + 1,
to find another rung that is lower, such that you can move to rung n + 1 from the
lower one.
If you can do these two things then clearly you can make it to any desired level.
This is what induction does: imagine that the n-th step of the ladder symbolizes
proving statement P (n). The “base case” means that you should check explicitly
the lowest n for which the statement P (n) makes sense, and the “crank” requires
you to provide a logical argument that says “If P (k) is true for all k ≤ n then
P (n + 1) is also true”. This “crank” is called the inductive step where the part “If
P (k) is true for all k ≤ n” is known as the inductive hypothesis. The “base case”
is the induction basis.
Remark I.1. In many cases, you will only use P (n) in order to prove P (n + 1),
but there are exceptions where using only P (n) is not convenient. Some people call
usage of all P (i) with i ≤ n “strong induction”. But there is nothing strong about
this sort of induction: one can show that what can be proved with strong induction
can also be proved if you just assume P (n) for the sake of proving P (n + 1).
Example I.2. We consider the question from the start of the section: show
that 0 + 1 + . . . + n = n(n + 1)/2. So, for n ∈ N we let the statement P (n) be
“0 + 1 + . . . + n = n(n + 1)/2”.
The base case would be n = 0 or n = 1, depending on your taste. In either
case the given statement is correct: if n = 0 then the sum on the left is the empty
sum (nothing is being added) and that means (by default) that the sum is zero. Of
course, so is 0(0 + 1)/2. One might be more sympathetic towards the case n = 1
in which the purported identity becomes 1 = 1(2)/2, clearly correct.
For the crank, one needs a way to convince other people that if one believes in
the equation
P (n) : 1 + 2 + . . . + n = n(n + 1)/2
P (n + 1) : 1 + 2 + . . . + n + (n + 1) = (n + 1)(n + 1 + 1)/2.
In induction proofs for equational statements like this it is usually best to compare
the left hand side (LHS) of the presumed and the desired equality and to show that
their difference (or quotient, as the case may be) is the same as those of the right
hand sides (RHS). In other words, one tries to manufacture the new equation from
the old.
In the case at hand, the difference of the LHSs is visibly n + 1. The RHS
difference is (n+1)(n+2)/2−n(n+1)/2 = (n+2−n)(n+1)/2 = 2(n+1)/2 = n+1.
So, if one believes in the equation given by P (n) then, upon adding n + 1 on both
sides, one is forced to admit that equation P (n + 1) must also be true. This
completes the crank and the principle of induction asserts now that all statements
P (n) are true, simply because P (0) is and because one can move from any P (n) to
the next “higher” one via the crank.
Remark I.3. For the functionality of induction it is imperative that both the
base case and the crank are in order. (It’s clear that without crank there is not
1. INDUCTION 15
much hope, but the checking of the base case is equally important, even if the crank
has already been established!)
Consider for example the following attempt of proving that 1 + 2 + . . . + n =
n(n+1)/2+6. Let’s write P 0 (n) to be the statement “1+2+. . .+n = n(n+1)/2+6”.
Now argue as follows: suppose that for some n ∈ N, P 0 (n) is true: 1 + 2 + . . . + n =
n(n + 1)/2 + 6. Add n + 1 on both sides to obtain 1 + 2 + . . . + n + (n + 1) =
n(n + 1)/2 + 6 + n + 1 = [n(n + 1) + 2(n + 1)]/2 + 6 = (n + 1)(n + 2)/2 + 6. So,
truth of P 0 (n) implies truth of P 0 (n + 1).
Of course, if you believe that we did the right thing in Example I.2 above, then
P 0 (n) can’t hold ever (unless you postulate 6 = 0). The problem with climbing the
P 0 -ladder is that while we have a crank that would move us from any step to the
next step up, we never ever actually are on any step: the base case failed!
Remark I.4. The usual principle of induction only works with collections of
statements that are labeled by the natural numbers. If your statements involve
labels that are not natural numbers then, typically, induction cannot be used in-
discriminately.
One can make various errors in induction proofs. Indicated here are two, by
way of an incorrect proof.
(1) “Theorem”: all horses have the same color.
Proof by induction: let P (n) (n ∈ N) be the statement “within any
group of n horses, all horses have the same color”. The base case P (0) is
void (there is no horse to talk about, so P (0) is true) and P (1) is clearly
true as well.
Now suppose P (n) is true and we prove P (n + 1) from that. It
means that we must show that in any group of n + 1 horses all horses
have the same color. So let S be a group of n + 1 horses, which we
name H1 , H2 , . . . , Hn+1 . Let T1 stand for the size n group of the first
n horses, T1 = {H1 , . . . , Hn }. Let T2 stand for the last n horses, T2 =
{H2 , . . . , Hn+1 }. Since T1 has n horses in it, statement P (n) kicks in and
says that all horses inside T1 have the same color, which we denote by c1 .
Similarly, all horses in group T2 (of size n) have all one color, called c2 .
However, the horses H2 , . . . , Hn appear in both sets, and so have colors
c1 and c2 simultaneously. We conclude c1 = c2 and so all horses in S had
the same color!
(2) “Theorem”: Let a be any positive real number. Then, for all n ∈ N, one
has an = 1.
Proof by induction: let P (n) be ”an = 1”. The base case is n = 0. In
that case, a0 = a1−1 = a1 /a1 = 1.
Now assume that P (i) is true for all 0, . . . , n.. We want to show
an
an+1 = 1. Rewrite: an+1 = an an−1 . Both “an = 1” (statement P (n)) and
“an−1 = 1” (statement P (n − 1)) are covered by the inductive hypothesis
and so P (n + 1) must be true.
Both proofs imply wrong results, so they can’t really be proofs. What’s the prob-
lem? The errors are not of the same type, although similar. In the first case, we
use the collection H2 , . . . , Hn of horses that are simultaneously in both T1 and T2 .
The problem is that if n = 1 then there aren’t any such horses: T1 is just {H1 } and
T2 is just {H2 }. So there is actually no horse that can be used to compare colors
in group T1 with those in group T2 , and so c1 and c2 have no reason to be equal.
16 I. WEEK 1: INTRODUCTION
In the second proof, you were sneakily made to believe that “an−1 = 1” is
covered by the inductive hypothesis, by calling it “P (n − 1)”. But if n = 0 then
n − 1 is negative, and we are not entitled to use negative indices on our statement!!!
One must be very careful not to feed values of n outside the set of naturals into an
inductive argument.
Remark I.6. One can formally turn induction “upside down”. The purpose
would be to prove that all statements in an infinite sequence P (n) are false. The
idea is: check that P (0) is false explicitly; then provide a crank that shows: if some
statement P (n) is true then there must already be an earlier statement P (i) with
i < n that is true.
This reverse method is called infinite descent and illustrated in the next exam-
ple.
Well-order of the natural numbers is a very basic property and closely related
to the following “obvious” result:
1. INDUCTION 17
size of n. So we have shown that if P (n) holds then n is even and P (n/2) also
holds.
In concrete terms, if the first n for which P (n) holds is called b then b must
be even and P (b/2) is also true. Of course, in most cases b/2 is less than b (so b
wouldn’t actually be the first), and the only natural number for which this does
not cause a problem is b = 0. So if there is any n with P (n) true, then P (0) should
also √be true. But, as we checked, it isn’t. So we deduce that P (n) is always false
and 2 must be irrational.
18 I. WEEK 1: INTRODUCTION
2. Arithmetic
We must begin with some algebra. We officially meet the definition of a ring
only in week X > 1, but I state it here already. The idea is to list all the important
properties of the set of integers.
2. ARITHMETIC 19
Output: gcd(a, b) = ci .
From what we said above, gcd(cj , cj+1 ) = gcd(cj−1 , cj ) at all stages of the
algorithm. In particular, gcd(a, b) = gcd(ci , ci+1 ) = gcd(ci , 0) by our choice of
aborting the loop. The gcd of any number and zero is that “any number”, so
gcd(a, b) is really the last nonzero remainder ci we found.
There is another aspect to the Euclidean algorithm, which is the following.
The last equation says how to write ci in terms of the previous two: ci = ri−1 =
ci−2 − qi−1 ci−1 . The second to last equation can be used to express ci−1 in terms
of ci−2 and ci−3 . Substituting, we can write ci in terms of ci−2 and ci−3 . Iterating
this backwards, one arrives at a linear combination of the form gcd(a, b) = αa + βb
for suitable integers α, β. This is a fact to remember:
Proposition I.18. Working backwards from the end of Euclid’s algorithm de-
termines a Z-linear combination
gcd(a, b) = αa + βb.
2. ARITHMETIC 21
we get c = αca + βcn = αac + βab = a(αc + βb). In particular, a divides c. But
then, the equation cn = ab becomes (c/a)an = ab and cancellation of a shows that
n divides b. Note that this part used the Euclidean algorithm, and is not true for
all rings.
Theorem I.24. Integers enjoy unique factorization. This means first off that
for all 0 6= n ∈ Z there is an prime factorization, which is an equation
n = c · p1 · · · pk
where each pi is a prime number and where c is a unit (which in Z implies c = ±1).
It means secondly that any two such factorizations are almost the same: if
d · q1 · · · q` = n = c · p1 · · · pk
are two such factorizations (c, d units and pi , qj prime) then k = ` and (up to sign
and suitable reordering) pi = qi
For example, 14 = 1 · 2 · 7 = (−1) · 2 · (−7) are two different but essentially
equivalent prime factorizations of 14.
Proof. What we need to show comes in two stages: given a natural number
n, we need to show it factors at all into prime factors. And then we need to show
that any two such factorizations agree, up to reordering. (Note that we can focus
on n > 0, since −n = (−1) · n and so a factorization of n corresponds to one of
−n).
We use strong induction. The base case is clear: 1 and 2 are surely factorizable
into units and primes: 1 = 1 and 2 = 2. So we focus on the crank. So let 2 ≤ n ∈ N
and assume that the numbers 1, 2, . . . , n all have a factorization into positive prime
numbers. (We don’t need to show that we can factor stuff into positive prime
numbers, but it is convenient when the number to be fctored is already positive).
We consider now n + 1. There are two cases: either n + 1 is prime, in which case
we can write n + 1 = 1 · (n + 1) as prime factorization. Or, n + 1 is not prime. Then
n + 1 is also not irreducible (since we will show later that prime = irreducible, not
prime = not irreducible), and so it factors as n + 1 = 1 · a · b with a, b not units.
Since n + 1 was positive, we can arrange a, b to be positive, and so they both fall
into the set of numbers 1, 2, . . . , n about which we already know that they can all
be factored. So, factor a = 1 · a1 · · · ak and b = 1 · b1 . . . b` into primes, so that
n + 1 = a · b = 1 · a1 · · · ak · b1 . . . b` has a factorization. So, all natural numbers do
have prime factorizations.
Now we need to show that these factorizations are unique. Take any natural
number n with two prime factorizations c1 · a1 · · · ak = n = c2 · b1 . . . b` where c1 , c2
are units and each ai , bj is a prime number. If any ai or bj is negativem we can
turn their signs by moving the signs into c1 and c2 . So all ai , bj can be assumed to
be positive.
Since a1 is prime, and since it divides the product c2 · b1 · · · b` , a1 must divide
one of the factors of this product. It cannot divide c1 since c1 = ±1 and a1 as a
prime has absolute value 2 or more. So, a1 divides some bt , so bt = a1 q1 for some
integer q1 . But bt was supposed to be prime, hence irreducible, so q1 is a unit. But
a1 and bt are positive, so q = 1 and a1 = bt .
Divide out a1 = bt to get c1 · a2 . . . ak = n/a1 = c2 · b1 · · · b̂t · b` , where the
hat indicates that bt has disappeared from the product. So these are two prime
2. ARITHMETIC 23
Exercise I.28. Show that the harmonic series diverges (has no finite value).
(Hint: what can you say about each ai ?).
Suppose you only took the terms that are inverses of prime numbers,
1 1 1 1 1 1 1 1 1 1 1
+ + + + + + + + + + + ··· ,
2 3 5 7 11 13 17 19 23 29 31
how does this series behave? We will need to use the following fact.
Exercise I.29. Show:
24 I. WEEK 1: INTRODUCTION
1−xn+1
(1) For any real or complex number x and any integer n, 1−x = 1+x+
x2 + · · · + xn . P∞ 1
(2) As long as |x| < 1, i=0 xi = 1−x .
Theorem I.30. The sum of the reciprocals of all the prime numbers is still
divergent.
This of course implies also that there are infinitely many primes. However, its
proof is far more delicate than Euclid’s proof above. We give here the idea behind
the proof; that the steps can be made rigorous is somewhat involved.
Proof. Any positive integer n is uniquely the product of positive prime num-
bers. The emphasis is on “unique”, if you multiply together different sets of numbers
you get different end products, pun intended.
Consider now the product
1 1 1 1 1 1 1 1 1
(1 + + + + . . .) · (1 + + 2 + . . .) · (1 + + 2 + . . .) · · · (1 + + 2 + . . .),
2 4 8 3 3 5 5 p p
where p is some prime number.
If you actually multiply this out, the resulting mess contains, for each choice
of finitely many primes p1 , . . . , pk bounded by p, the quantity 1/(p1 · · · pk ). So, the
mess actually contains exactly one copy of the inverse of every natural number that
has a prime factorization in which only powers of primes occur that are bounded
by p. Taking the limit p → ∞, one might try to believe in an equation
∞
X 1 Y
= (1 + 1/p + 1/p2 + 1/p3 + . . .),
n=1
n
p prime
and conclude that the right hand side is, like the harmonic series on the left, infinite.
The art (which we omit) is to make this argument and all that builds on it, so that
it become mathematically sound. P∞
Using the geometric series, this suggests n=1 n1 = p prime 1−1/p 1
Q
. Now take
1
P P
logs on both sides. The RHS turns into p prime ln( 1−1/p ) = − p prime ln(1 −
1/p). Looking at the P graph of the log-function, one sees that ln(1 − x) ≤ x, so
1 1
P
p prime ln( 1−1/p ) ≤ p prime p . But the left hand side was already infinite, so
therefore the sum of the prime reciprocals must be too.
This theorem says that there are still quite a bit of primes, namely enough
to make the sum diverge. (Remember: if you just looked at the subsum given
by powers of your favorite prime, this would be geometric and hence convergent).
How many primes are there in comparison to all numbers? This is best asked in
the context of prime density.
Theorem I.31 (Prime Number Theorem). Let pk be the k-th prime number.
pk
Then the fraction k·ln(k) approaches 1 as k → ∞.
Equivalently, if you pick randomly a number n near the number N then the
chance that n is prime is about ln(N )/N .
Another set of questions one could ask is about primes in arithmetic progres-
sions (rather than in all numbers). That means: let
A(a, n) = {a, a + n, a + 2n, . . .}
2. ARITHMETIC 25
Here, the input s is to be taken a complex rather than a real number. This sum
will converge (absolutely) if the real value of s is greater than 1 because of things
we know about the geometric series from Exercise I.29. On the other hand, the
harmonic series teaches that at s = 1 the value of ζ is infinite. In the places where
s has real part less than 1, one can use a graduate-level technique called “analytic
continuation” to make sense of the series (even though it probably diverges). The
result is a nice function in s that can have poles every now and then (such as in
s = 1). Values of the zeta function appear in physics (how odd!) and chemistry
(no more even!), and they have a tendency to involve the number π. At negative
even integers, ζ(s) is zero for reasons that come from a “functional equation” that
ζ satisfies:
ζ(s) = 2s π s−1 sin(πs/2)Γ(1 − s)ζ(1 − s).
Here, π is the usual π from trigonometry, and Γ is a version of “factorial” for non-
integer input. (If you believe this equation, you must believe that ζ(−2n) = 0 by
looking at the contribution of the sine).
26 I. WEEK 1: INTRODUCTION
3. Modular arithmetic
We are perfectly used to claim that 4 hours after it was 11 o’clock it will be
3 o’clock. In effect, we equate 12 with zero in these calculations. In this section
we learn how to calculate on more general “clocks” and even solve equations on
“clocks”.
3.1. Computing “modulo”: Z/nZ.
Definition I.38. For any integer n, write nZ for the set of all integer multiples
of n, so nZ stands for {. . . , −3n, −2n, −n, 0, n, 2n, . . .}.
For an integer a write then a + nZ for the collection of all integers who leave
remainder a when divided by n, so a + nZ = {. . . , a − 2n, a − n, a, a + n, a + 2n, . . .}.
Note that a + nZ = (a + n) + nZ. Such sets we call cosets modulo n, while
the various integers that float around in a given coset are called representatives.
They are also sometimes written as “a mod n”. If the value of n is understood
from the context, we may write a for a + nZ. (If n = 12, and if a = 3, then
3 = 3 mod 12 = 3 + 12Z = {. . . , −21, −9, 3, 15, 27, . . .}. This is the set of all times
on an absolute clock at which a usual clock shows “3o’clock”).
Finally, write Z/nZ (“zee modulo enn zee”) for the collection of all cosets
modulo n. (If n = 12, this is the set of all possible full hours, clustered by what a
12-hour-clock would show).
Here is an example on a very small “clock”.
Example I.39. Let n = 4. there are four cosets modulo 4, namely
(1) 0 + 4Z = {. . . , −8, −4, 0, 4, 8, . . .}, 1 + 4Z = {. . . , −7, −3, 1, 5, 9, . . .},
(2) 2 + 4Z = {. . . , −6, −2, 2, 6, 10, . . .}, 3 + 4Z = {. . . , −5, −1, 3, 7, 11, . . .}.
So, Z/4Z = {0, 1, 2, 3}. Representatives of 3 include 3, 7, −133 among others.
Remark I.40. While this “modular arithmetic” may seem a bit esoteric, be
assured that it is far more important than you can imagine. For example, all
computers on this planet calculate in Z/2Z or a slightly more complicated scheme.
Without modular arithmetic, there would be no twitter, no email, no instagram.
Not even a digital watch.
3. MODULAR ARITHMETIC 27
Amusingly, one can calculate with these cosets as if they were numbers. Re-
garding addition, we always knew that 4 hours after it was 11 o’clock it will be 3
o’clock because the coset of 11 plus the coset of 4 gives the coset of 15, which is to
say, of 3. That one can also multiply is a new idea:
(a + nZ) + (b + nZ) := (a + b) + nZ;
(a + nZ) − (b + nZ) := (a − b) + nZ;
(a + nZ) · (b + nZ) := (a · b) + nZ.
The amusing part is that this works well on any choice of representatives. For
example, in order to compute (2 + 7Z) + (3 + 7Z) you could say: pick representative
−5 = 2 + (−1) · 7 for 2 and representative 24 = 3 + 3 · 7 for 3. Add them to get 19
and so (2 + 7Z) + (3 + 7Z) = 19 + 7Z. Of course, you probably would have chosen
2 = 2 + 0 · 7 and 3 = 3 + 0 · 7 as representatives, resulting in the coset of 5. The
point is that 5 and 19 are actually the same. In order to prove that this is always
ok and not just in our explicit example you should carry out
Exercise I.41. Show that for all choices of a, a0 , b, b0 , n with n|(a − a0 ) and
n|(b − b0 ) one has:
• n divides (a + b) − (a0 + b0 ); (this says that the cosets of a + b and of a0 + b0
always agree);
• n divides (a − b) − (a0 − b0 ); (this says that the cosets of a − b and of a0 − b0
always agree);
• n divides ab−a0 b0 ; (this says that the cosets of ab and of a0 b0 always agree).
3.2. Divisibility tests. Suppose you are asked “is 1234567 divisible by 3?”.
You could sit down and calculate, or ask a friend with a computer, but you could
also think. Such as: n = 1234567 comes to me as a decimally expanded number,
n = 10k · ak + · · · + 10 · a1 + a0 where k = 6, a6 = 1, a5 = 2, a4 = 3, a3 = 4, a2 = 5,
a1 = 6 and a0 = 7. In order to test divisibility of n by 3, I’d like to know whether
n mod 3 is zero or not. But, n mod 3 = (10k · ak + · · · + 10 · a1 + a0 ) mod 3, and
“ mod ” goes well over addition and multiplication:
X
n mod 3 = (ai mod 3) · (10 mod 3)i
X
= (ai mod 3) · (1 mod 3)i
X
= ai mod 3.
It follows that n is a multiple of 3 if and only if the sum of its digits is a multiple
of 3. Of course, if you want, you can reapply this idea to the output:
1234567 mod 3 = (1 + 2 + 3 + 4 + 5 + 6 + 7) mod 3 = 28 mod 3
= (2 + 8) mod 3 = 10 mod 3
= (1 + 0) mod 3 = 1 mod 3.
Hence, 1234567 leaves rest 1 when divided by 3.
Obviously, a similar thought works for 9 instead of 3 since any power of 10
leaves rest 1 when divided by 9.
Exercise I.42. Prove that 11 divides n if and only if it divides a0 − a1 + a2 −
a3 + · · · + (−1)k ak .
28 I. WEEK 1: INTRODUCTION
Question: Following the single arrows just counts how big the current digit is.
What is the function of the double arrows?
CHAPTER II
Week 2: Groups
1. Symmetries
Example II.4. Let OS be the oriented square . Its symmetry group has
fewer elements than that of the square, namely only the rotations {e, `, `2 , `3 } with
29
30 II. WEEK 2: GROUPS
a composition table
e ` `2 `3
` `2 `3 e
`2 `3 e `
`3 e ` `2
since `4 = e and there is no other relation. This group of symmetries is called the
cyclic group C4 , since you only need to know one element of it (such as `) and
everyone else is a power of it. The “4” comes from the fact that `4 = e and no
lower positive power will do (or, because there are 4 elements in this cyclic group.
Itis like a clock with 4 hours).
Example II.5. Now we look at the symmetries of the letter H. It has 4 elements,
the identity e, the rotation x by 180◦ the left-right flip ↔ and the up-down flip l.
The table is
e ↔ l x
↔ e x l
.
l x e ↔
x l ↔ e
(You should actually check a few of the products listed here).
This set of symmetries is called the Klein 4-goup and denoted KV4 . Felix Klein
was the superstar of symmetry investigations. Note that Sym(H) ⊆ Sym() since
drawing a box H around the H does not change the symmetries.
Note also that the tables for KV4 and C4 are seriously different since e shows
up on the diagonal with different multiplicity. (The element e is special and can be
recognized even if you use different letter as it is the one element for which ex = x
for every symmetry x).
2. Groups
We are now ready to define what a group is. It generalizes the symmetry studies
above.
Definition II.6. A group is a set G with an operation · that takes ordered
pairs (g, g 0 ) from G × G and “multiplies” them to other elements of G. (In other
symbols, · : G × G → G). This operation must satisfy the following conditions:
(1) a · (b · c) = (a · b) · c∀a, b, c ∈ G (associativity);
(2) there is an identity or neutral element e ∈ G such that ∀g ∈ G one has
e · g = g · e = g;
(3) ∀g ∈ G there is an inverse element g̃ ∈ G with g · g̃ = g̃g = e.
Remark II.7. (1) As a matter of convenience we often skip the dot and
just write ab for a · b (as we have done above for symmetries). Also, one usually
writes g −1 for the g̃ in item (3) of the definition.
(2) Be warned, that one of the conditions you might have placed here is missing:
we do not require that ab = ba in general. If you think of group elements as
prodecures that you compose, this is clear: it makes usually some difference whether
you put on socks first and then shoes, or the other way round.
(3) Note the following quirk of this asymmetry: if ab = c then c−1 = b−1 a−1 .
Thinking of socks and shoes makes this more obvious. You also have seen this for
taking inverses of matrices, and of course a matrix is just a procedure acting on a
2. GROUPS 31
vector (by multiplication), so this all fits together. The invertible n × n matrices
with real entries are one of the standard examples of a group. It is called the general
linear group Gl(n, R).
(4) Associativity implies that the product g · g · · · g of many copies of the same
element is uniquely defined and does not depend on in what order we multiply the
copis. (For example, you could take 4 copies and multiply them like ((gg)g)g or
like ((gg)(gg)). For 3 factors, this is explicit from the associativity rule, and for
more than 3 we discuss it in Lemma II.13 below).
Theorem II.8. The symmetries on any chosen object form a group.
Proof. The set G is the set of symmetries, the operation of G is composition
of symmetries. The identity is the symmetry that does not move. The inverse of a
symmetry is the symmetry done backwards. The associativity rule comes from the
fact that it holds for composition of functions (where it boils down to reading the
definition of composition).
To each group one can write down a table similar to the tables we have looked at
for symmetries. For group tables one uses the phrase Cayley table. You surely have
noticed at this point that each row and column of such table contains each element
(once). That is no accident:if you had ac = bc for example, the same element
showing up as product twice in the same column, then also (ac)c−1 = (bc)c−1 and
so a(cc−1 ) = b(cc−1 ), or ae = be which entails a = b according to the various group
axioms. We say that groups have the cancellation property.
Example II.9. Here is a list of important groups with their operations. The ∗
just indicates usual multiplication.
(1) (Z, +), the integers with addition.
(2) (Z/nZ, +), modular addition (verification in HW);
(3) (Rn , +), the n-dimensional vector space has as part of its axioms the group
axioms for +;
(4) ({1, −1}, ∗) with a Cayley table similar to that of the dihedral group D2 ;
(5) (R>0 , ∗), which contains the previous group and uses the same operation,
identity and inverse;
(6) (R\{0}, ∗), which contains the previous group and uses the same operation,
identity and inverse;
(7) (Gl(n, R), ∗) and (Gl(n, C), ∗) as previously mentioned;
Example II.10. We consider here the list of all possible groups with 4 or fewer
elements.
(1) If |G| = 1 then G is just e and the Cayley table is that of the dihedral
group D1 .
(2) If |G| = 2 then G = {e, f } for some other element f , and by the cancellation
rule f f can’t be ef and so must be e. So G has a Cayley table essentially that of
the dihedral group D2 .
(3) If |G| = 3, G = {e, a, b}. Since ab can’t be ae = a, but also not eb = b, by
cancellation, it must be ab = e. Then we are forced to concede aa = b and bb = a,
and so a3 = e. So the table is the one you get from the rotational symmetries of
e a a2
the equilateral triangle alone: a a2 e , with b = a2 . This is essentially C3 .
a2 e a
32 II. WEEK 2: GROUPS
associativity. For higher powers, argue as follows. Suppose we have proved that
the product of k copies of g is independent of the placement of parentheses. Then,
for i + j = k + 1 and i, j > 0 we have (g i )(g j ) = (g i )(g(g j−1 )) = ((g i )g)(g j−1 ) =
(g i+1 )(g j−1 ). So one may shuffle one copy of g after the other from one factor to the
other without changing the product. So, a product of k copies of g only depends
on g and k but not the placing of parentheses.
Let a, b be in a cyclic group generated by g. According to the definition of
a cyclic group, there are numbers i, j ∈ Z such that a = g i , b = g j . But then
g i g j = g j g i since they are both the product of i + j copies of g.
Definition II.14. If in a group (G, ·) it is true that gh = hg for all g, h ∈ G
then G is Abelian.
Cyclic groups are Abelian, but lots of groups are not, such as Sym( ). (the
elements a, b, c only have two different powers, e and themselves, and `, r only have
the three powers e, `, r). Also, Sym(H) is not cyclic as one sees easily from the
squares.
The question when a power of an element is e seems to be important:
Definition II.15. For an element g of the group (G, ·), the smallest number
k ∈ N>0 such that g k = e is its order ord(g). (There might not be such a k (like
for 3 ∈ (Z, +) for example. We then say ord(g) = 0 or ord(g) = ∞.)
We call |G| the order of the group.
Inside Sym( ), both the powers of ` and the powers of a form what we call a
subgroup.
Definition II.16. If (G, ·) is a group, then a subgroup is a subset H ⊆ G
which, when equipped with the multiplication of G, is a group in its own right.
If you take an element g ∈ G then the order of g in G is the same as the order
of its twin in G0 . This follows from the translation of the Cayley tables. Basically,
this says that if φ is the bijection then φ(a ·G b) = φ(a) ·G0 φ(b).
The upshot is that one can use order to discriminate between groups. For
example, KV4 is not C4 because KV4 has 3 elements of order 2, and C4 only one.
One can also count subgroups and compare: KV4 has 5 subgroups, namely
{e}, {e, ↔}, {e, l}, {e, x}, KV4 . But C4 has only three: {e}, {e, `2 }, C4 . So these
two groups cannot be isomorphic.
Recall that for sets A, B the Cartesian product A × B is the set of all ordered
pairs (a, b) with a ∈ A, b ∈ B.
Definition II.19. If G, G are groups, then G × G0 is also a group, with mul-
tiplication (g1 , g10 ) · (g2 , g20 ) = (g1 ·G g2 , g10 ·G0 g20 ).
For example, (R2 , +) is simply (R, +) × (R, +).
Example II.20. The cyclic groups C2 = {e, a} with a2 = e and C3 = {e, b, b2 }
with b3 = e have Cayley tables as discussed earlier. In these groups, e has order 1,
a has order 2 and b has order 3. What about elements of C2 × C3 ?
The list of elements has 2×3 members, and they are (e, e), (e, b), (e, b2 ), (a, e), (a, b), (a, b2 ).
One sees easily that (e, e) has order 1 = lcm(1, 1); (e, b) and (e, b2 ) have order
3 = lcm(1, 3); (a, e) has order 2 = lcm(2, 1); and (a, b) and (a, b2 ) have order
6 = lcm(2, 3).
(We explain the lcm statements: in general one has ord(x, y) = lcm(ord(x), ord(y)).
Why? Surely, the lcm is a power that sends (x, y) to (e, e). The powers satisfy
xk = xord(x)+k and y k = y ord(y)+k . So y k = e implies y k = y ord y = e and so
y gcd(ord(y),k) = e. But the gcd can’t be bigger than ord(y) because it needs to
divide it, and it can’t be smaller than the order because of the definition of order.
The only way out is that the order is the gcd. So the order of y divides any k with
y k = e. Similarly, the order of x divides any exponent i with xi = e and the order
of (x, y) divides any exponent with (xi , y i ) = (e, e). So whatever the order of (x, y)
is, it must be a multiple of ord(x) and ord(y), while being as small as possible.
That is simply the lcm.
4. Automorphisms
Definition II.21. An automorphism of a group G is a a relabeling of its
elements that preserves the Cayley table.
For example, C3 is the group {e, a, b} with rules ab = ba = e, ea = ae = a, be =
eb = b. This is completely symmetric in a, b. So the bijection
a 7→ b,
b 7→ a,
e 7→ e
is an automorphism of C3 . (Geometrically, this switches left roation with right
rotation in the rotational symmetries of an isosceles triangle).
In principle, an automorphism is just a special permutation of the elements.
So one can search through all permutations and just keep those that preserve the
Cayley table. This is not efficient if G has many elements, one should use the group
structure in the search.
4. AUTOMORPHISMS 35
Note, that one possible automorphism is always just to leave everything as is.
That is like the e in a group. In fact, automorphisms do form a group in their
own right. e ∈ Aut(G) is the relabeling that sends every element of G to itself;
multiplication of two automorphisms is just doing one relabeling after the other;
the inverse of an automorphism is the relabeling done backwards.
Looking at C3 : there are only two automorphisms, the identity on C3 , and the
switch discussed above. This is because eG must be sent to eG (yx = x for all x is
something only the element y = e does, and relabelings must preserve products!).
Composing the switch with itself gives the identity on C3 . So, it is fair to say that
Aut(C3 ) is basically the group with table as in Example II.10 part (2).
Another interesting example occurs in C4 , which is the group of symmetries of
the oriented square , with elements a = `, b = `2 , c = `3 and the understanding
4
` = e. Here one can interchange a and c while keeping e, b fixed:
a 7→ c,
c 7→ a,
b 7→ b,
e 7→ e.
It is easy to check that this relabeling preserves the table when written with a, b, c, e.
Again, this is the only automorphism since we must send e to e (the only
element of order one) and b to b (the only element of order two). Aside, of course,
of the identity on C4 leaving everything fixed. So, Aut(C4 ) is the “same” group as
Aut(C3 ), both isomorphic to C2 , sameness in the sense that ther Cayley tables are
the same after renaming.
Example II.22. The automorphisms of KV4 are more interesting.
Suppose we fix a. Then we could fix b but that also forces us to fix c as the only
remaining element of order 2. That then comes down to the identity, sending each
element of KV4 to itself. Alternatively, if we do not fix b, the only open destination
for b is c. So a 7→ a, b 7→ c, c 7→ b.
Alternatively, we can try sending a 7→ b. If we fix c then we are in a similar
situation as before, because then b must go to a. On the other hand, we could send
a 7→ b and b 7→ c which forces c 7→ a.
The cases where a 7→ c are similar, with the letters b and c exchanged.
Altogether, the 6 options are summed up in the following table, where each
row represents an automorphism, and where it sends the elements of G is recorded
in the row.
e a b c
ψe e a b c
ψa e a c b
ψb e c b a
ψc e b a c
ψ` e b c a
ψr e c a b
For notation: ψe keeps everyone fixed; ψx for x ∈ {a, b, c} keeps e, x fixed and
switches the other two; ψ` encodes a rotation (b, c, a) of the letters a, b, c to the left
in the sense that we read the sequence (b, c, a) as the instruction a goes where b
36 II. WEEK 2: GROUPS
was, b goes where c was, and c goes where a was, which is now really a rotation to
the left), and ψr moves them according to the instruction the right to make (c, a, b).
The notation is intentionally reminding you of Sym( ). Indeed, if you align
ψx in Aut(KV4 ) with x ∈ Sym( ) then you find this to be an isomorphism (see
Definition IV.7 below): it is a one-to-one correspondence between the elements
of Sym( ) and the elements of KV4 . For example, ψ` after ψa first sends a
to a, and then to b. And it sends b first to c and then that c is sent to a. So
ψ` ψa is e 7→ e, a 7→ b, b 7→ a, c 7→ c. This is the same effect as that of psic ,
and so ψ φa = ψc . If we compare to the Cayley table of Sym( ) then we
also have correspondingly `a = c. Checking the entire list of products, we see
Aut(KV4 ) = Sym( ).
5. Free groups
Definition II.23. A group is free (on the elements g1 , . . . , gk ) if, for some
k ∈ N, it is isomorphic to the group Fk of all words in the letter set Lk =
{e, x1 , . . . , xk , y1 , . . . , yk } with the rules (and no other rules) of
• ez = z = ze for all z ∈ L;
• xi yi = e = yi xi for all 1 ≤ i ≤ k;
• associativity.
Here, the group operation is simply writing two words next to each other in the
given sequence.
These groups are “free” because their elements have no other constraints aside
from the group axioms. They are not Abelian for k > 1 (since we do not require
xi xj = xj xi ). In contrast, F1 = {. . . , y12 , y1 , e, x1 , x21 , . . .} is isomorphic to the
Abelian group (Z, +) via the identificatiuon xk1 ↔ k ∈ Z, y1k ↔ −k ∈ Z.
There are also free groups on infinite numbers of letter. We will not look at
them much.
It is a fact that all subgroups of a free group are free (basically, because there
are no relations, but the proof is not so easy), and somewhat shockingly, F2 contains
subgroups isomorphis to F3 , F4 , . . .. We won’t discuss this phenomenon.
It is also a fact that one can take any group G and interpret it as a free group
“with extra rules”.
Definition II.24. If G is a group we call a list L of elements a generating set
if every element of G is a product of elements from L ∪ L0 where L0 is the list of
inverses of L.
If such list has been chosen, we refer to elements of L as generators.
Evidently, L = G is a generating set, although usually not an interesting one.
Example II.25. Z × Z is generated by {(1, 0), (0, 1)}. Because of this we can
view Z × Z as “F2 with the additional rules x1 x2 = x2 x1 and x1 y2 = y2 x1 and
x2 y1 = y1 x2 and y1 y2 = y2 y1 .
To se this note first that we get the relations x1 y1 = y1 x1 and x2 y2 = y2 x1 for
free, because all four of these products give e.
Secondly, we read x1 as (1, 0), and x2 as (0, 1), which then suggests y1 is (−1, 0)
and y2 is (0, −1). Then all additional rules imposed on F2 above correspond to Z×Z
being Abelian.
CHAPTER III
The main hero in this week is the group Z/nZ with addition, where n ∈ N.
Recall that it is a cyclic group, generated by the coset of 1. The order of the element
1 + nZ is n as one easily sees, and the order of Z/nZ is also n.
All groups Z/nZ are Abelian, because Z is Abelian and we just install new
rules in order to make Z/nZ from Z.
a, b ∈ Z. Note that this gcd is positive since n, i are. Then compute g an+bi =
(g n )a (g i )b = ea eb = ee = e. So gcd(n, i) is an exponent that when used over g gives
e. But n = ord(g) is supposedly the smallest positive exponent of this sort. So,
gcd(n, i) = n and so n|i.
For the last part, g i = g j implies, when multiplying witht he inverse of g j , that
i−j
g = e, which then by the first part gives n|(i − j). If on the other hand we have
n|(i − j) then g i−j = e and so g i = g j .
Definition III.3. If g ∈ G we write hgi for the group of al powers—negative
and positive—of g in G. This is the cyclic subgroup generated by g.
Corollary III.4. Up to renaming, (hgi, ·) is (Z/ ord(g)Z, +) in the sense that
the renaming identifies the Cayley tables.
Proof. Let n = ord(g). Then we associate to g i ∈ hgi the element 1 + nZ in
Z/nZ. Then g i · g j = g i+j corresponds to (i + nZ) + (j + nZ) = (i + j) + nZ, and
g n = e to (1 + nZ) + . . . + (1 + nZ).
| {z }
n copies
The next result then tells us how the groups generated by powers of g ∈ G will
look.
Corollary III.5. Let g ∈ Ghave order n. Then the group hg k i generated by g k
is the same group as the group hg gcd(n,k) i that is generated by g gcd(n,k) . Moreover,
abstractly this group is the same as the cyclic group Cn/ gcd(n,k) .
Proof. By the same argument as in the previous proof, hg k i contains g gcd(n,k) ,
and so also all its powers. Conversely, gcd(n, k) divides k and so of course hg gcd(n,k) i
contains g k and all its powers. So, the groups hg k i and hg gcd(n,g) i are contained
one in the other in both directions and hence equal.
Let h = g gcd(n,k) . What could the order of h be? Write n = d · gcd(n, k); then
h = (g gcd(n,k) )d = g n = e and so the order of h is no more than d. But if hi = e
d
for some i < d then we also have e = (hi ) = (g gcd(n,k) )i , and this would contradict
ord(g) = n since gcd(n, k) · i < gcd(n, k) · d = n.
We can now complete a table from above on subgroups of Z/12Z:
g := k mod 12Z size of hgi gcd(n, k) n/ gcd(k, n) = ord(k + 12Z)
1, 5, 7, 11 12 1 12
2, 10 6 2 6
3, 9 4 3 4
4, 8 3 4 3
6 2 6 2
0 1 12 1
Looking at this table, the next natural question is: how do you predict the
exponents i that give an equation hgi = hg i i?
As a starter, let’s ask for the generators of Z/nZ, the guys for which hgi is the
entire group Z/nZ. For n = 12, the relevant cosets are 1, 5, 7, 11. These are the
numbers that are coprime to 12. (Note: any representative in k is coprime to n when
k is coprime to n. For example, gcd(5, 12) = 1 and so also gcd(5 + 127 · 12, 12) = 1
and 5 + 127 · 12 lives in the same coset as 5.)
The magic therefore lies in coprimeness.
2. PRODUCTS AND SIMULTANEOUS MODULAR EQUATIONS 39
Definition III.6. For n ∈ Z let φ(n) be the Euler φ-functin that counts the
number of cosets in Z/nZ that consist of representatives coprime to n.
For example, φ(12) = 4 since modulo 12 the cosets 1+12Z, 5+12Z, 7+12Z, 11+
12Z are those that are made of numbers coprime to 12.
Lemma III.7. If G = hgi is cyclic of order n then the generators of G are
exactly the elements g k with gcd(n, k) = 1.
Proof. Any element h of G is some power h = g k of g since G = hgi. A
generator is an element g k of G with hg k i = G, which is the case exactly when
ord(g k ) = n. But ord(g k ) = n/ gcd(n, k), and so we find that g k is a generator if
and only if gcd(n, k) = 1. So counting the generators is the same as counting the
cosets of Z/nZ that are mnade of numbers coprime to n.
We can now move and ask when hg i i = hg j i for some exponents i, j. Since the
size of hg i i is n/ gcd(n, i) we find the implication
[hg i i = hg j i] ⇒ [gcd(n, i) = gcd(n, j)].
In reverse, if the gcd equality holds, then gcd(n, i) = gcd(n, j) is a divisor of j
which forces g j inside hg gcd(n,i) i and so hg i i = hg gcd(n,i) i contains hg j i. Exchenging
i, j gives the reverse containment, hence an equality.
We have now seen all parts of
Theorem III.8. Let g be an element of order n. So hgi is Cn up to relabeling.
(1) Subgroups of cyclic groups are always cyclic.
(2) For all i ∈ Z, ord(g i ) divides n and equals n/ gcd(n, i).
(3) If k|n then there is a unique subgroup of size k inside hgi, and it is exactly
the set of n/k-th powers hg n/k i of g n/k .
(4) If k|n then he number of elements of order k inside hgi is equal to φk. If
k 6 |n, no elements have order k.
(5) Obviously, if g i generates a subgroup of order k then it does not generate
a subgroup P of order different from k. It follows from the previous item,
that n = d|n φ(d).
To see the last part in action, look at Z/nZ. Our table above on elements and
groups they generate runs in the left column through all the cosets and puts them
into one row if they generate the same subgroup. There are 12 such elements, they
get grouped as
12 = |{z}
4 + |{z}
2 + |{z}
2 + |{z}
2 + |{z}
1 + |{z}
1 .
=φ(12) =φ(6) =φ(4) =φ(3) =φ(2) =φ(1)
Proof. If the gcd condition is in force, take the element g = (1, . . . , 1). Its
order is a multiple of every ni , but as they have no common factor, it is a multiple
of the product n1 · · · nk , which is |G|. But no element can have order greater than
|G|, so ord(g) = n1 · · · nk and so g generates G.
On the other hand, any element of G is always of order at most lcm(n1 , . . . , nk ),
since this power creates the neutral element in very component of the product. If
gcd(ni , nj ) > 1 for any i 6= j then this lcm cannot be the product n1 ·n2 · · · nk = |G|,
so everyones order is less than |G|. So G will then have no element of order |G|.
In particular, this says that a product Z/(p1 )e1 Z × Z/(p2 )e2 Z × . . . × Z/(pk )ek Z
for distinct primes p1 < p2 < . . . < pk is cyclic.
Note that our first example showed that distinctness is crucial.
Example III.11. Let’s try to make this more explicit. We know that Z/7Z ×
Z/5Z is cyclic, and must be of order 7×5 = 35. So abstractly we know Z/7Z×Z/5Z
is Z/35Z in disguise. But can we see that inside Z/35Z?
We are looking for an identification of Z/35Z with the product Z/7Z×Z/5Z that
preserves the Cayley table (which means it has to preserve the group operation +).
Let’s make a naı̈ve guess: take i+35Z and attach to it the element (i+7Z, i+5Z) in
Z/7Z×Z/5Z. Surely, this attachment will respect addition since (i+35Z)+(j+35Z)
would be attached to (i + 7Z, i + 5Z) + (j + 7Z, j + 5Z) = ((i + j) + 7Z, (i + j) + 5Z)
as you would expect. We write π for this recipe, π(i + 35Z) = (i + 7Z, i + 5Z).
(Important note here: in Z/35Z, we have grouped numbers together into a coset
whenever they differ by a multiple of 35. Since multiples of 35 are also multiples
of both 5 and 7, we can make “cosets of cosets” and read for example the cosets
3 + 35Z, 8 + 35Z, 13 + 35Z, 18 + 35Z, 23 + 35Z, 28 + 35Z, 33 + 35Z as a partition of
the coset 3 + 5Z in Z/5Z. So, moving from i + 35Z to i + 5Z actually makes sense
since it does not destroy cosets but preserves them and makes them even larger. So
it is actually legal to go from Z/35Z to Z/5Z by the assignment “i + 35Z becomes
i + 5Z”. Same argument for going from Z/35Z to Z/7Z. But you could not, for
example, go from Z/35Z to Z/6Z: in Z/35Z, 3 and 38 belong to the same coset,
but in Z/6Z they do not. Destroying cosets isnot legal when moving groups about.)
So we have a way to go from Z/35Z to Z/7Z × Z/5Z. Big question, how do we
go back? In other words, given a pair (a + 7Z, b + 5Z) in Z/7Z × Z/5Z, how do we
find i + 35Z such that (a + 7Z, b + 5Z) = π(i + 35Z)?
What we know is that this is supposed to work based on the fact that 5 and 7
are coprime. So gcd(7, 5) = 1 must get used somewhere. The Euclidean algorithm
says that there are numbers x, y ∈ Z with 1 = 7x + 5y. (Specifically, x = −2 and
y = 3 works). Then let’s consider the number i = a · y · 5 + b · x · 7. (That one
should look at this is not obvious and only becomes clear after a good number of
examples). Then we compute:
(a · y · 5 + b · x · 7) + 7Z = a · y · 5 + 7Z = a(1 − 7x) + 7Z = a + 7Z,
(a · y · 5 + b · x · 7) + 5Z = b · x · 7 + 5Z = b(1 − 5y) + 5Z = b + 5Z.
We have basically proved:
Lemma III.12. If m, n are relatively prime and a, b ∈ N are given, then the
simultaneous equations
i mod mZ = a mod mZ,
i mod nZ = b mod nZ
3. U (n): AUTOMORPHISMS OF Z/nZ 41
Again, if you have 3 or more coprime numbers, one gets a corresponding results
on products of unit groups.
Example III.18. How many elements does U (750) have?
The bad way is to write them all out. The enlightened 453 student says:
750 = 3 · 53 · 2, and so U (750) = U (3) × U (53 ) × U (2). I know |U (3)| = 2,
|U (53 )| = 53−1 (5 − 1), and |U (2)| = 1. Hence |U (750)| = (2) · (25 · 4) · 1 = 200.
Remark III.19. Recall the Euler φ-function that counts for n ∈ N how many
numbers from 1, . . . , n are relatively prime to n. Recall also that a is relatively
prime to n if and only if a + nZ is a generator of the group Z/nZ. (In other words,
the order of a + nZ is n, or yet in other words, na is the lowest positive multiple of
a that is divisible by n).
Since U (n) is made of the cosets of Z/nZ that come from numbers relatively
prime to n, there are exactly φ(n) elements in U (n). That means also that if m, n
are relatively prime, then φ(mn) = φ(m)φ(n) because of the theorem above.
CHAPTER IV
1. Equivalence relations
Definition IV.1. Let S be a set. An equivalence relation is a binary relation
' on S such that
• a ' a for all a ∈ S (reflexivity);
• [a ' b] ⇔ [b ' a] for all a, b ∈ S (symmetry);
• [a ' b and b ' c] ⇒ [a ' c] for all a, b, c ∈ S (transitivity).
Examples of such equivalence relations are :
• the usual equality of numbers
• congruence of geometric figures;
• equality in the module caculation (this is really the relation i ' j on Z
whenever n|(i − j).
An example of a realtion that is not an equivalence relation is the usual ≤,
because it is not symmetric: 3 ≤ 4 but not 4 ≤ 3.
Lemma IV.2. If S is a set with equivalence relation ' then one can partition
S into cosets/equivalence classes where any coset contains all the elements that are
mutually equivalent to one another.
If we denote the cosets S1 , S2 , . . . , then we have: Si ∩Sj is empty unless Si = Sj .
Moreover, S is the union of all Si .
Lemma IV.3. Let G be any group, and pick n ∈ N. Then let A be the collection
of all group elements a ∈ G that have order exactly n. Then |A| is a multiple of
φ(n).
Proof. If a ∈ A, then hai is a cyclic group of order n. By last week’s results,
hai contains exactly φ(n) elements whose order is exactly n, and these are exactly
the generators of hai. So, make an equivalence relation on A where x ' y if and
only if hxi = hyi. Each equivalence class has size φ(n), they do not meet, and their
union is A. So the coset size divides |A|.
Note: if G is cyclic and n = |G|, then G = hgi and |A| = φ(n) by last week.
2. Morphisms
0
Let G, G be two groups.
Definition IV.4. A morphism (or homomorphism) is a function ψ : G → G0
that respects the group operations:
ψ(g1 ·G g2 ) = ψ(g1 ) ·G0 ψ(g2 )
for all g1 , g2 ∈ G.
45
46 IV. WEEK 4: COSETS AND MORPHISMS
We saw last week that the automorphisms of Z/nZ are labelled by the cosets
k + nZ with gcd(k, n) = 1. In other words,
Aut(Z/nZ, +) = U (n).
Example IV.9. While U (12) and U (10) both have 4 elements, they are not
isomorphic. Indeed, U (10) is cyclic generated by 3, and U (12) has no element of
order 4.
One natural way of making automorphisms is the following:
Definition IV.10. Let a ∈ G be a group element. Define a map ψa : G → G
by setting
ψa (g) = aga−1 .
This is the inner automorphism on G induced by a.
Note first that this indeed respects multiplication: ψa (g1 )ψa (g2 ) = ag1 a−1 ag2 a−1 =
ag1 g2 a−1 = ψa (g1 g2 ).
Note next that if G is Abelian, then ψa (g) = aga−1 = aa−1 g = ea = a for any
choice of g and a. So in an Abelian group, every inner automorphism is the identity
map.
We will see later that since Inn(Sym( )) has 6 different elements just like
Sym( ) itself, then the conclusion is that innG is actually an isomorphism, so that
as abstract groups there is no difference between Sym( ) and Inn(Sym( )).
Note that p prime is essential: Fermat’s little theorem fails for p = 4. Question:
are there non-primes for whch this does work?
4. KERNELS AND NORMAL SUBGROUPS 49
in the kernel of φ. So we test it: φ(x ·G g ·G x−1 ) = φ(x) ·G0 φ(g) ·G0 φ(x−1 ) =
φ(x) ·G0 eG0 ·G0 φ(x−1 ) = φ(x) ·G0 φ(x−1 ) = eG0 . So, indeed we have xgx−1 ∈ H. So
H is stable under conjugation.
Definition IV.24. If H ⊆ G is a subgroup that is stable under all inner
automorhisms, we call H a normal subgroup.
As a side remark, this is not “normal” behavior in the usual sense of language.
Looking at all subgroups H of a given group G, it is usually quite unnormal for H
to be normal. Normal subgroups are quite special.
Note that aHa−1 = H is equivalent to aH = Ha so that left and right cosets
agree for each a ∈ G precisely when H is normal.
Example IV.25. The kernel of any morphism is normal as we proved in the
theorem above.
Example IV.26. The subgroup {e, a} ⊆ Sym( ) is not normal. Indeed,
a ∈ H but `a`−1 = c`−1 = cr = b is not in H.
One can check that there are not many normal subgroups of Sym( ): the only
ones stable unde all conjugations are the trivial group {e}, the rotation subgroup,
and the whole group. (These latter two are always normal and never interesting as
subgroups).
50 IV. WEEK 4: COSETS AND MORPHISMS
Remark IV.27. A subgroup is normal if and only if the left cosets aH agree
with the right cosets Ha. This follows because [aH = Ha] ⇔ [aHa−1 = H] as one
sees by multiplying with a−1 on the right.
Example IV.28. Let G be the 2 × 2 invertible matrices with real entries, with
matrix multiplication as group operation. Let φ : G → R× be the morphism that
takes determinants. Linear algebra says that det(ABA−1 ) = det(A) det(B)/ det(A) =
det(B). So if det(B) = 1 then this is also true for all its conjugates.
CHAPTER V
Definition V.3. There is another way to wrote permutatins called cycle no-
tation. You start with some letter (A,
and then record where A goes. For example,
A B C
for the right rotation r = we write down (A, C. Next you ask where
C A B
C goes, and under r that is B. So we continue to (A, C, B. But B now is moved
to A and that “closes the cycle, so we write (A, B, C).
If a cycle closes before you have witten down
what happens to all
elements,
A B C D E F
just open another cycle. So, the permutation has cycle
B C A D F E
notation (A, B, C)(D)(E, F ). It rotates A, B, C in a 3-cycle and also rotates E, F
in a 2-cycle, and leaves D put.
One may or may not indicate 1-cycles (since they are talking about elements
that do not move, the assumption is that if an element does not show in a cycle
that you wrote, then it is not moving. For example (1, 3, 5) is a permutation that
leaves 2 and 4 fixed.)
A cycle of length 2 is a transposition.
How does one compose cycles? Just the same as always: start on the right.
So, (1, 4, 5)(2, 3, 4, 1)(3, 5) is decoded as follows. Start with 1. under (3,5) it goes
to position 1, then 1 goes under (2,3,4,1) to position 2. So the 1 we started with is
now in position 2. Stuff in position 2 moves under (1,4,5) not al all, so position 2
is the final destination of 1. So we start writing the prouct as (1, 2.
Next we redo this all with input 2. Under (3,5), 2 stays put. Under (2,3,4,1)
stuff in bucket 2 moves to bucket 3. And then under (1,4,5) stuff in bucket 3 stays
put. So overall, 2 moves to bucket 3. So we are now at (1, 2, 3.
Restart with input 3. Under (3,5), 3 moves to bucket 5, and under (2,3,4,1)
bucket 5 stays put. The at the end (1,4,5) mive bucket 5 to bucket 1, and that
means our 3 lands in bucket 1. So, we have found the first part of the product cycle
as (1,2,3).
This does not yet explain what happens to 4 and 5 under the product. Let’s
check 4. Under (3,5) the bucket 4 stays put. Then it is moved to bucket 1 under
(2,3,4,1). And bucket 1 is moved to bucket 4 under (1,4,5). Hence the number 4
stays put overall. That means, 5 also must stay put since there is no more open
space. So, the product is (1, 2, 3)(4)(5).
Remark V.4. • What our product procedure produces is disjoint cy-
cles. That is, the cycles we write down as anser are such that no number
occurs in more than one cycle. Disjoint cycles are prefereable since we
“understand” better.
• For example, the order of any cycle (in the group theory sense) is its own
length: if you rotate left a bunch of k people on a round table, you need
to repeat this k times until everyone is back to his own seat. Moreover,
if you have the product of a bunch of disjoint cycles, then the order of
this product is th lcm of the cycle lengths. For example, the order of
(1,2,3)(4,5) is 6, because only iteration multiples of 3 make 1,2,3 go back
home, and only even numbers of iterations make 4,5 go back home.
In contrast, (1,2,3)(2,3,4)=(1,3)(2,4) has order 2, not 3 (the lack of
disjointness messes with thing!)
Theorem V.5. Any permutation is a product of 2-cycles (usually not disjoint!).
V. WEEK 5: PERMUTATIONS AND THE SYMMETRIC GROUP 53
Proof. It is enough to show that any single cycle can be made from 2-cycles.
If n = 2 this is clear (except that you need to say that the identity is writeable as
(1, 2)(1, 2). )
If n ≥ 3 check that (a1 , . . . , ak ) = (a1 , a3 , . . . , ak )(a1 , a2 ). So the theorem
follows from induction.
Lemma V.6. If you take a permutation σ and write it as product of permuta-
tions, then the number of transposition is not determined by σ, but the number of
permutations for the same σ is either always odd or always even.
Before we embark on a proof, one more concept:
Definition V.7. Let σ be a permutation of 1, . . . , n. We say that [i, j] is a
switch of σ if i < j but σplaces i in a position behind where it places j. In other
1 2 ... n
woords, if you write σ = then i < j but σi > σj .
σ1 σ2 . . . σ n
The disorder of σ is the number of switches of σ. The parity of σ is the answer
to the question “Is the disorder of σ even or odd’ ?”
1 2 3 4
For example, the cycle (1,2,3,4) is the permutation and so has
2 3 4 1
switches [1, 4], [2, 4], [3, 4] and so has disorder 3 and is an odd permutation (has odd
parity)
Proof of the lemma: It is enough to prove that the identity cannot be written
as the product of an odd number of transpositions. (Since e = (1, 2)(1, 2) is an even
way of writing the identity).
The main idea is that if some σ is composed with a transposition then its
for 1 ≤ i < j ≤ n,
parity changes. Let’s check that. So we imagine, composing the
1 2 ... n
transposition (i, j) with the permutation σ = and we count
σ1 σ2 . . . σn
the change in the disorder.
Let’s say the output of σ is the sequence
s1 , . . . , si−1 , si , si+1 , . . . , sj−1 sj , sj+1 , . . . , sn .
Then the output of (i, j)σ is
s1 , . . . , si−1 , sj , si+1 , . . . , sj−1 si , sj+1 , . . . , sn .
We consider the change in the number of switches.
If a switch of σ does not involve i nor j then it is a switch also of the composition
(i, j)σ. So we need to focus on switches that involve either i or j or both. We next
study when a switch involves one of i, j.
If k < i, the number of st that appear to the right of sk but are smaller than
sk does not change if we interchange si ith sj .
If k > j then the number of st that are to the right of sk and are smaller than
sk chenges even less.
If i < k < j, there are 4 cases:
(1) If sk < si and sk < sj then [k, i] is a switch in σ and [k, j] is a switch in
(i, j)σ.
(2) If sk < si and sk > sj then [k, i], [k, j] are a switch in σ and neither of
[k, i], [k, j] is a switch in (i, j)σ.
54 V. WEEK 5: PERMUTATIONS AND THE SYMMETRIC GROUP
(3) If sk > si and sk < sj then neither of [k, i], [k, j] is a switch in σ and both
[k, i], [k, j] are a switch in (i, j)σ.
(4) If sk > si and sk > sj then [k, j] is a switch in σ and [k, i] is a switch in
(i, j)σ.
In all cases then, so far, the change of the number of switches in σ versus (i, j)σ is
even.
Finally, consider the pair i, j. If it is not a swutch for σ then it must be one for
(i, j)σ, and conversely. So overall, the number of switches is an even number plus
one, and hence odd.
What this means is that if you write any σ as product of transpositions, the
number of transpositions must agree with the parity of σ (which does not depend
on how you write σ as such product!). So, the number of transpositions used is
even if and only of the parity of σ is even.
Example V.11. Recall that KV4 is the symmetry group of the letter H. We
can make it a subgroup of S4 as follows. Take as symbols of the group the “letters”
e, ↔, l, x. Now ask what the effect of multiplying by the group elements is on the
sequence {e, ↔, l, x}. We find
e · {e, ↔, l, x} = {ee, e ↔, e l, e x} = {e, ↔, l, x}
l ·{e, ↔, l, x} = {l, x, e, ↔}
↔ ·{e, ↔, l, x} = {↔, e, x, l}
x ·{e, ↔, l, x} = {x, l, ↔, e}.
V. WEEK 5: PERMUTATIONS AND THE SYMMETRIC GROUP 55
If one now reads these as permutations, one can write them as cycles as:
e becomes ()
l becomes (e, l)(↔, x)
↔ becomes (e, ↔)(l, x)
x becomes (e, x)(l, ↔).
If one translates into symbols 1, 2, 3, 4 we get
KV4 = {(), (1, 2)(3, 4), (1, 3)(2, 4), (1, 4)(2, 3)}.
Remark V.12. In many cases, there is a more obvious way of embedding a
given group into a symmetric group. For example, the symmetry group of a cube
is a naturally a subgroup of S8 since the symmetries of the cube move around the
8 vertices. But that is sort of an accident: not every group comes to us as the
symmetry group of a small set of things (with certain constraints). If someone
hands us the symmetry group of a cube without saying what it really is, and if
we don’t notice it, we would have to take recourse to the recipe of the proof of
the proposition. And that would view the symmetry group of the cube (with 48
elements) as a subgroup of S48 , a rather unpleasant idea. So the proposition conveys
a principle, but it pays to be opportunistic.
CHAPTER VI
Let me start with recalling some ideas from the past. If H ⊆ G is a subgroup
(same identity element, same multiplication) then is is a normal subgroup if it has
no conjugate subgroups aside form itself. This is saying, that aHa−1 = H (or
aH = Ha)for any a ∈ G. Note that this says that aha−1 is again in H, but it does
not require that aha−1 = h (but it also does not say this should not be true)
Let us also recall that H can be used to make H-formed clusters in G by looking
at the left cosets aH; any element of G belongs to one such coset, so their union is
all of G, and two cosets either do not meet at all, or agree completely. No partial
agreement is possible (because of the cancellation property).
1. Making quotients
Definition VI.1. Let us denote the collection of all left H-cosets in G by
G/H.
Note the similarities: when G = Z is all integers, and H = nZ the subgroup of
integers divisible by n then G/H = Z/nZ is exactly the collection of cosets a + nZ
with a ∈ Z.
Note also that Z/nZ is a group itself; we would like to arrange for G/H to be a
group as well. The natural plan would be to define (aH) ∗ (bH) = abH. Lets look
in an example what that is like.
Example VI.2. Let G = S4 be the symmetry group of the equilateral tetra-
hedron (also known as the permutation on 4 elements) and take as H the group of
permutations {(), (12)(34), (13)(24), (14)(23)}. We saw at the end of last class that
this group can be identified with KV4 , the symmetry group of the letter H.
As S4 has 24 elements and H has 4, the clusters we make for cosets have size
4 and there will be 6 such cosets. They are: (no shortcut here, I just sat down and
computed each set aH by hand):
E := H = {(), (1, 2)(3, 4), (1, 3)(2, 4), (1, 4)(2, 3)},
γ := (12)H = {(1, 2), (34), (1, 3, 2, 4), (1, 4, 2, 3)},
β := (13)H = {(1, 3), (1, 2, 3, 4), (2, 4), (1, 4, 2, 3)},
α := (14)H = {(1, 4), (1, 2, 4, 3), (1, 3, 2, 4), (2, 3)},
λ := (123)H = {(1, 2, 3), 1, 3, 4), (2, 4, 3), (1, 4, 2)},
ρ := (124)H = {(1, 2, 4), (1, 4, 3), (1, 3, 2), (2, 3, 4)}.
Now we would like to make these 6 clusters into a group. As mentioed above, we
aim for (aH)(bH) = abH. In order to avoid problems such as we met in Assignment
4a when we were looking at morphisms from Z/mZ to Z/nZ that were not even
functions (because they destroyed cosets), we need to keep cosets together. More
57
58 VI. WEEK 6: QUOTIENTS AND THE ISOMORPHISM THEOREM
e ↔ E, a ↔ α, b ↔ β, c ↔ γ, r ↔ ρ, ` ↔ λ,
we see that up to the renaming we are looking at Sym( ) = S3 .
Could we have seen this somehow? Yes, I think so, and here is how. The fact
that we have any group structure at all on G/H is because the normality of H
assures us that whenever a0 belongs to the coset aH and b0 belongs to the coset bH,
then a0 b0 is in the coset abH. Now look at the 6 cosets, and pick out the elements
in each coset that do not use 4. We find E ∈ E, (1, 2) ∈ γ, (1, 3) ∈ β, (2, 3) ∈ α,
1. MAKING QUOTIENTS 59
(1, 2, 3) ∈ λ and (1, 3, 2) ∈ ρ. The remarkable fact is that there is exactly one in
each coset. Composing or inverting these elements can only produce other elements
that also do not use 4, so these 6 elements actually form a group by themselves,
a subgroup of S4 . And it is easy to see that this subgroup is exactly S3 . The
renaming was made in such a way that a Greek letter corresponds to the Roman
letter that we have the element of S3 sitting inside the coset indicated by the Greek
letter.
Let us look at a somewhat easier example, easier because of commutativity.
Example VI.3. Let G = (Z/24Z, +) and let H be the subgroup formed by
the multiples of 6. As in the previous example, |G| = 24 and |H| = 6. But in this
case there is no question that H is normal, since G is Abelian and so even ah = ha
element by element, and not just aH = Ha as a set.
The cosets are then
(3) {0 + 24Z, 6 + 24Z, 12 + 24Z, 18 + 24Z},
(4) {1 + 24Z, 7 + 24Z, 13 + 24Z, 19 + 24Z},
(5) {2 + 24Z, 8 + 24Z, 14 + 24Z, 20 + 24Z},
(6) {3 + 24Z, 9 + 24Z, 15 + 24Z, 21 + 24Z},
(7) {4 + 24Z, 10 + 24Z, 16 + 24Z, 22 + 24Z},
(8) {5 + 24Z, 11 + 24Z, 17 + 24Z, 23 + 24Z}.
So, for example, the last of these cosets contains all numbers that leave rest 5 when
divided by 6. If we call these collections 0, . . . , 5 (in the given order), and recall
that we are supposed to use addition, it is clear how we want to think of the group
G/H: it is Z/6Z.
We formulate officially what we have seen in examples.
Theorem VI.4. If H is a normal subgroup of G then one can equip the col-
lection of left cosets {aH|a ∈ G} with a group structure. The multiplication in this
group takes aH and bH and multiplies them to abH. The resulting group is denoted
G/H and called the quotient group of G by H.
If H is normal, the same construction can be carried out for the right cosets
Ha, and that also leads to a group. One can check that these are the same groups,
so that the symbol “G/H” is unambiguous.
It is often good for understanding a definition when one sees a case where the
defined concept is absent.
Before we state it, let me remind you that you have seen something like this
before: if A is a real m × n matrix, you can view it as a way to turn vectors v ∈ Rn
into vectors A · v of Rm . Both Rn , Rn are groups (the first three axioms that you
learned for a vector space mean just that it is a group for addition of vectors!)
The kernel of the matrix A used to be the vectors v that have Av = 0, and since
the zero vector is the identity element for vector addition, this old “kernel” idea
for vector spaces agrees exactly with our new one for groups. And you were also
told that the image of A (you used to call it the column span) is a vector space
(hence group!) of dimension rk(A). And to top it all off, you learned that rank lus
nullity gives n. In new and fancy terms this can be phrased as “the kernel of A is a
vector space of dimension n − rk(A), and Rn / ker(A) is a vector space of dimension
n − (n − rk(A)) = rk(A). This quotient is precisely the column space of A, a vector
space of dimension rk(A) just like Rn / ker(A).
Theorem VI.7 (The isomorphism theorem). If
ψ : G → G0
is a morphism of groups with kernel H := ker(ψ) sitting inside G, and with image
im(ψ) sitting inside G0 , then there is an isomorphism
ψ : G/ ker(ψ) ' im(ψ)
where ψ(aH) = ψ(a).
Here, the group operation in G/H is (aH)(bH) = abH and the operation in
im(ψ) is the one from G0 .
I will not prove this theorem in detail, but here is why you should think it is
true:
(1) As you move from G to G0 using ψ, products are preserved but all of H
is crunched dwon to eG0 , basically by definition. Therefore if you want
to relate stuff n G with stuff in G0 , you need to form the cosets G/H to
account for the “lumping together” of anything in H.
(2) You are not going to be able to relate elements of G0 that are not outputs
of ψ to anything in G since ψ is your only comparison vehicle, and stuff
in G0 that ψ “does not see” is stuff that ψ has no opinion about.
(3) So, really the question is what G/ ker(ψ) hasto do with im(ψ). And the
function φ I mentioned, which sends a coset aH to ψ(a), can be shown
to be a morphism (easy, since ψ is), and injective (confusing, but easy) ,
and surjective (easy). But that makes it an isomorphism.
In particular, if ψ is surjective,
G/ ker(ψ) ' im(ψ).
Example VI.8. Here I will talk through some examples.
(1) Let ψ be the morphism from Z/15Z to Z/25Z that multiplies by 5, sending
a + 15Z to 5a + 25Z. (Recall that if k is to be used as morphism from
Z/mZ to Z/nZ then we need that n|km).
Then ker(ψ) = {a+15Z|5a+25Z = 0+25Z}. This requires that 25|5a
so that a must be a multiple of 5. So, ker(ψ) = {0+15Z, 5+15Z, 10+15Z}.
You can view this as Z/3Z “inflated by a factor of 5”.
The image im(ψ) of ψ are the cosets {5a + 25Z}. That is a group of
5 elements. We know (5 is prime) that this is a cyclic group, and indeed
62 VI. WEEK 6: QUOTIENTS AND THE ISOMORPHISM THEOREM
This says that the transformation T (which exists independently of the choice of
W V
basis) is represented relative to the bases B 0 , B 0 by the matrix A0 = QW AQ−1 V .
(As a special case, if V = W and one chooses BV = BW then the change of
coordinates has the effect of conjugation on A. In some sense this is clear: if you
have a recipe tor a tranformation (called A) that works in one language (the bases
BV , BW ) and you want to use in a different language (the bases BV0 , BW
0
) then you
first translate the ingredients from the new into the old language (by Q−1 V ), then
use the recipe (namely A), and then translate the result into the new language (by
QW ). Once again, this goes right to left because that is the way functions work).
The moral of this linear algebra story is that a transformation is not affected
by the way we think of the input and the output space, but the tools we use
63
64 VII. WEEK 7: FINITELY GENERATED ABELIAN GROUPS
Row reduction says that the relations of these 4 columns don’t change (and
neither does therow span) if you add row 1 to row 3 the 1 to wipe out the -1,
1 4 0 3
which leads to 0 3 9 12. Of course, it does have an effect on the column
0 3 3 6
span of the matrix, so this amounts to a coordinate change in Z3 (the target of the
map).
Now our rowreduction can go on, with 3 as pivot, erasing 3 below it. We get
1 4 0 3
0 3 9 12 . That is another change of basis in the target space Z3 . We
0 0 −6 −6
can now change the -6 to a 6, and then normal row reduction would stop.
The row steps are a reflection of a change of basis in the target of the transfor-
mation, but we can also change basis in the source. That is encoded by (invertible)
column operations. For example,
we can use the top left 1 to wipe out the other
1 0 0 0
numbers in row I to get 0 3 9 12. And now we can use the 3 to wipe out
0 0 6 6
1 0 0 0
all that is to its right: 0 3 0 0. And then the left 6 to kill the right 6:
0 0 6 6
1 0 0 0
0 3 0 0.
0 0 6 0
Definition VII.2. The shape of this matrix is called Smith normal form. It
features: only nonzero entries on the diagonal, and from upper left to lower right
the diagonal entries are multiples of the previous entries.
The business of base change in source and target does not change the structure
of the quotient group (target/rowspan), although it changes how we think of it (as
any coordinate change does). So, our quotient group G/H now turns out to be Z3
modulo the linear combinations of the columns of the last matrix above. In other
words,
The point of the row reduction work is that the stuff in H now has been “decoupled”:
the first coordinate of an element of H is any number, the second is any number
divisible by 3, the last is any multple of 6. The coordinates do no longer “talk to
each other”, they have become independent.
This also makes clear what G/H is equal to: (Z/1Z) × (Z/3Z) × (Z/6Z). Note
that Z/1Z is the trivial group as 1Z = Z.
Recall, that Z/6Z = (Z/2Z) × (Z/3Z) since 2, 3 are coprime. So G/H =
(trivial group) × (Z/3Z) × ((Z/2Z) × (Z/3Z)).
There is one big hurdle we did not meet in the previous example: our pivots
came as a free gift. The following example shows what to do when lunch is not
free.
66 VII. WEEK 7: FINITELY GENERATED ABELIAN GROUPS
3
Example VII.3. Lets try this for H the subgroup of G = Z generated by
10 −6 4
(10, −4, 8)T , (−6, −6, −16)T , (4, −10, −8)T , which yield the matrix −4 −6 −10.
8 −16 −8
There is no element here that can be used as a pivot, because a pivot should divide
all the other numbers it is used to wipe out (we don’t have access to fractions. . . ).
This means, we have to make a pivot first, by clever row or column operations, or
both.
The main question is what we can hope and aim for. Surely, we can’t make a 1
here since all numbers are even. But we could hope for a 2, and that would divide
every other number. And we can make a 2 by subtracting row III from row
I, to
2 10 12 2 10 12
get −4 −6 −10. Now clean out the front column: 0 14 14 . Then
8 −16 −8 0 −56 −56
2 10 12
one more row step leads to 0 14 14 and then 3 column operations produce
0 0 0
2 0 0
0 14 0.
0 0 0
We infer that G/H ' (Z/2Z)×(Z/14Z)×(Z/0Z) = (Z/2Z)×(Z/2Z)×(Z/7Z)×
(Z) since 2, 7 are coprime.
Note that the zero on the diagonal is actually very important here, it tells us
about Z being a factor of G/H (and so makes G/H have infinitely many elements).
Definition VII.4. If A is an integer m × n matrix with m ≤ n then let A0
be the Smith normal form of A. The diagonal elements of A0 are the elementary
divisors of A.
If m > n, first augment A with m − n columns of zeros on the right, and then
proceed to compute Smith normal form. (This has the effect of adding m − n zeros
to the set of elementary divisors).
Theorem VII.5 (FTFGAG, Part 1). We assume that A is m × n, with m ≤ n.
If m > n, augment A with m − n columns of zeros. We start with properties of
Smith normal form and elementary divisors.
(1) The Smith normal form of A can be computed by row and column opera-
tions of types I and II.
(2) The Smith normal form of A is determined by A alone, and not on how
we compute the normal form by pivot choices.
(3) The elementary divisors d1 , . . . , dn of A are the m numbers on the diagonal
of the Smith normal form A0 of A.
(4) The elementary divisors satisfy di |di+1 ∀i.
2. Generating groups
Recall that if a group Q is cyclic with generator g then the elements of H are
powers of either g or g −1 . This gives a presentation
Q = Z/ ord(g)Z
as a quotient of Z by a suitable subgroup.
2. GENERATING GROUPS 67
1 2 0 0 1 2 0 0
0 −2 2 0. We then use the −2 as pivot to get 0 −2 2 0. At last, we
0 2 0 2 0 0 2 2
1 0 0 0
do column operations to get to 0 2 0 0, which certifies that KV4 = Z3 /H
0 0 2 0
is isomorphic to (Z/1Z) × (Z/2Z) × (Z/2Z) = (Z/2Z) × (Z/2Z). We can associate
KV4 with (Z/2Z) × (Z/2Z) via
e l ↔ x
(0 + 2Z, 0 + 2Z) (1 + 2Z, 0 + 2Z) (0 + 2Z, 1 + 2Z) (1 + 2Z, 1 + 2Z)
and this assignment is an isomorphism.
One can use this result to count the number of Abelian groups of a certain
order, and also compare different Abelian groups for being isomorphic.
Example VII.8. How many Abelian groups G with 168 elements are there?
For each, find the elementary divisors.
168 = 23 · 31 · 71 . By FTFGAG, G should be a product of some Z/2di Z
and Z/3ei Z and Z/7fi Z. Of course, in order to make the group indeed have 168
elements, we need the sum of the di to be 3, and the sum of the ei to be 1 and
the sum of the fi to be 1 as well. That actually leaves very little choice, since an
exponent of 0 can be ignored. We must have one e and one f of value 1. The only
interesting bit is how we partition 3. As we know, this could be as 1 + 1 + 1 or as
1 + 2 or as 3.
So the possibilities are:
(Z/8Z) × (Z/3Z) × (Z/7Z),
(Z/2Z) × (Z/4Z) × (Z/3Z) × (Z/7Z),
(Z/2Z) × (Z/2Z) × (Z/2Z) × (Z/3Z) × (Z/7Z).
The elementary divisors satisfy: the product is 168, they divide each other (and
1’s can be ignored since they lead to Z/1Z factors which are trivial). Since 3 and
7 appear with power 1 in 168, both appear only in the last (biggest) elementary
divisor. The possibilities are: 168, or 2 · 84, or 2 · 2 · 42. One can see that the
partitions of the exponent 3 of 2 correspond to these factorizations: 1 + 1 + 1
corresponds to (21 ) · (21 ) · (21 · 31 · 71 ), 1 + 2 to (21 ) · (22 · 31 · 71 ), and 3 to (23 · 31 · 71 ).
Of course, the same applies to the partitions of the exponent 1 over 3 and 7, but
since 1 can’t be partitioned nontrivially, that is not so thrilling and 3, 7 only appear
in the last elementary divisor.
We want to explain lastly, why for example G = (Z/2Z) × (Z/4Z) is not iso-
morphic to Z/8Z, and in the process understand all similar questions with more or
higher exponents.
The underlying reason is by finding elements that are “killed” by 2 (or its
powers) in this case. By this we mean elements g ∈ G that when you double them
are zero. In Z/2e Z, there is always exactly one element that is not zero but yet
killed by 2, namely the coset of 2e−1 . More generally, we learned when we studied
cyclic groups, that the number of elements in a cyclic group Z/nZ of exact order
d is either zero (when d does not divide n) or (if d|n) equals φ(d), the Euler phi
2. GENERATING GROUPS 69
function that counts numbers relatively prime to d. Since φ(2) = 1 this agrees with
the above search.
So in a cyclic group of order divided by p, the number of elements that are
killed by the prime number p is exactly φ(p) + 1, the 1 coming from the fact that
the identity is killed by p but already dead (and so did not count for the order-p-
count in φ(p)). But φ(p) + 1 = p, so in a cyclic group of order divided by p there
are exactly p elements killed by p if p is prime.
Now, in a product such as (Z/2Z) × (Z/4Z) there are now 2 × 2 elements killed
by 2, because if a pair is killed by 2 then each component is killed by 2. And since
there are 2 choices in each component, then there are 2 × 2 such pairs.
More generally, in a product (Z/pe1 Z) × · · · × (Z/pek Z), pk elements will be
killed by p. So, groups with a different number of factors of the sort Z/pe Z cannot
be isomorphic, because they have different numbers of elements that are killed by
p.
If the number of such Z/pe Z is the same, consider the number of elements
killed by p2 . In each Z/pei , if ei = 1 there are p elements killed by p2 , but if
ei > 1 then there are p2 such elements. So in (Z/pe1 Z) × · · · × (Z/pek Z) there are
p#{ei ≥1} · p#{ei >1} elements killed by p2 . So, groups with equal number of factors
of type Z/pZ but different numbers of factors Z/p2 Z are not isomorphic.
In this manner one can prove by induction the last part of the theorem.
Remark VII.9. The above is relevant to the finite part of a group (the part
to elementary divisors different from 0). In homework you will show that Zm and
Zn are isomorphic exactly if m = n. That then finishes the last part of FTFGAG
stated next.
Theorem VII.10 (FTFGAG, Part 3). Let G be any finitely generated Abelian
group, and choose generators g1 , . . . , gm . Then G has a presentation π : Zm G
with π(ei ) = gi where ei is the i-th unit vector of Zm . Then this identifies G =
Zm /H as Zm modulo some subgroup H of Zm . Here the
One can find a matrix A whose column span is exactly H. The elementary
divisors of A do not depend on the chosen presentation of G nor do they depend on
the chosen matrix A. They only depend on G.
The finitely generated Abelian group G is characterized by the elementary divi-
sors in the sense that two groups have the same elementary divisors if and only if
they are isomorphic.
CHAPTER VIII
We have seen in two different places that one can read a group as a bunch of
permutations. First, as symmetries of actual objects (like an equilateral triangle,
for example) where the permutations occur at special places of the objects (the
corners of the triangle). Secondly, and much more formally, we have interpreted a
group element g ∈ G as a permutation σ g of the elements of G via left mutiplica-
tion: σ g (g 0 ) = gg 0 . In this section we formalize this sort of idea and discuss some
consequences.
Definition VIII.1. Let X be a set and let G be a group. Under the following
circumstances we shall speak of a left action of G on X:
(1) There should be a way of “multiplying” any element of G onto any element
of X. In other words, we need a function
λ : G × X → X,
(g, x) 7→ λ(g, x).
We then want that this action behaves well with respect to group multi-
plication as follows.
(2) The identity element e = eG should ”fix” every element of X, so that we
have
λ(eG , x) = x
for all x ∈ X.
(3) Given any two group elements g, g 0 ∈ G we require
λ(g, λ(g 0 , x)) = λ(gg 0 , x).
We will look exclusively at left actions, and henceforth just say “action” when
we mean “left action”. (Just to fill the void: a right action ρ : X × G → X would
want that ρ(g 0 , ρ(g, x)) = ρ(gg 0 , x); note the reversion in the order of g, g 0 here).
We will often write less officially gx for the result λ(g, x) of g acting on x. Then
the two rules above become
eg = g ∀x ∈ X, ∀g ∈ G,
g(g 0 x) = (gg 0 )x ∀g, g 0 ∈ G, ∀x ∈ X.
I recommend thinking of the elements of X as physical objects (“points”) that
one can draw and touch, and the process λ(g, −) as a way of moving the points in
X about. Here, λ(g, −) is the process the lets g ∈ G act on all points of X, the −
is just a place holder.
I order to say interesting things about group actions, we need a few more
concepts that arise naturally.
Definition VIII.2. Let λ be an action of G on X and choose x ∈ X.
71
72 VIII. WEEK 8: GROUP ACTIONS
• The orbit of x is those points y in X that you can “get to from x” using
multiplication of x by elements of G. In symbols, denoting the orbit of x
by orbG (x),
orbG (x) = {y ∈ X|∃g ∈ G with gx = λ(g, x) = y}
|{z}
or simply orbG (x) = Gx.
• If starting from x, the action can carry you to all other points of X, then
we say that the action is transitive. If G acts transitively on X then it is
customary to call X a homogeneous G-space.
• Complementary to the orbit of x is the notion of the stabilizer of x,
StabG (x) = {g ∈ G with gx = x},
the group elements that do not move x. Here we say that g moves x if
gx 6= x.
• If no element of G moves x, that is when StabG (x) = G, we call x a fixed
point of G. If g does not move x, we say that x is a fixed point for g, or
that g fixes x. We write FixX (g) for the points x ∈ X for which gx = x.
Remark VIII.3. You will show in homework that StabG (x) is a subgroup of
G.
We consider some examples, concrete and abstract.
Example VIII.4. Let G = Sym( ) and let X consist of the vertices of the
triangle. As we said many times, G can also be interpreted as SX , the permutation
group on the elements of X.
Let x be the A-vertex. Then StabG (x) consists of the identity e and the A-flip
a, since the other 4 elements b, c, `, r of G all move x.
Similarly, the stabilizer of C is {e, c} and that of B is {e, b}.
The rotations `, r have no fixed points, and the fixed points of e are al points
of X. The reflections a, b, c have only one fixed point each.
The action is transitive, since already the rotations are capable to carry any
point to any other point.
Example VIII.5. Let G be the symmetris of a cube, and let G be the rigid
symmetry group of a cube. cube. (This is the subgroup of all symmetries of the
cube consisiting of just the rigid motions that are cube symmetries). We found that
|G| = 48, 24 rotations from G, plus 24 non-rigid motions that are a composition of
a rotation and the antipodal map (which sends each vertex to the one diametrically
across).
Let X be the vertices of the cube and study the action of G (or G) on X.
If x is the upper left front vertex, there are 3 rigid motions that stabilize it (the
3 rotations that fix the big diagonal on which x lies) and then 3 more non-rigid
motions that combine the antipodal map with the 3 rotations that exchange x with
its antipode. So | StabG (x)| = 3 and StabG (x)| = 6.
Both actions are transitive. (Since G ⊆ G, it is enough to check that for G,
but we know that one can rotate any vertex of the cube into any other).
Most elements of G have no fixed point in X. Note that if a motion fixes a
vertex, it must also fix the antipodal point of that vertex. The 2 × 4 non-trivial
rotations that fix a big diagonal have 2 fixed points. The identity of G has 8 fixed
points.
VIII. WEEK 8: GROUP ACTIONS 73
The 3×4 motions from G that combine the antipodal map with a rotation about
one of the big diagonals followed by a reflection about the plane perpendicular to
this diagonal also have two fixed points.
Example VIII.6. Let G be any group and H a subgroup. We do not require
H to be normal. Let G/H be the set of all cosets gH relative to H. We take X to
be G/H and act on it by left multiplication:
λ(g, g 0 H) = gg 0 H for all g, g 0 ∈ G.
It is straightforward to check the group action rules: λ(eG , gH) = eG gH = gH, and
λ(g, λ(g 0 , g”H)) = gg 0 g”H = λ(gg 0 , g”H) because of associativity of multiplication
in H.
The stabilizer of a coset gH is the set of all a ∈ G with agH = gH, which says
g −1 agH = H and that is equivalent to g −1 ag ∈ H. For example, if g = e and so
gH = eH = H, the condition becomes a ∈ H, so the stabilizer of the “point” eH
in X is exactly H. In general, the the equation agH = gH means that for every
h ∈ H the expression agh should be of the form gh0 for some h0 ∈ H. That means
ag = gh0 h−1 and so a = gh0 h−1 g −1 . Since the product h0 h−1 is again in H, we find
that a must be ing gHg −1 . On the other hand, (gHg −1 )(gH) = gH(gg −1 )H =
gHH = gH so that the stabilizer of gH is exactly the set gHg −1 . This says that
the stabilizers of gH are always conjugate subgroups of H. In particular, if H
happens to be normal (but only then), each stabilizer is equal to H.
If gH wants to be a fixed point for multiplication by g 0 then we need g 0 gH = gH,
which amounts to g −1 g 0 gH = H. This forces g −1 g 0 g to be in H, so there should
be an element h ∈ H with g −1 g 0 g = h, or g 0 = ghg −1 . So, gH is a fixed point for
g 0 precisely if g 0 is in the conjugate subgroup gHg −1 .
In reverse, given g 0 then gH is fixed under multilication with g 0 precisely when
gHg −1 contains g 0 . Note that belonging to gHg −1 may not be very easy for g 0 . For
example, if H is normal then the condition “g 0 should belong to some conjugate
subgroup of H” just boils down to “g 0 must be in H”. Specifically, this applies in
an Abelian group Gm as then all subgroups H are normal.
We are interested in counting. That usualy means, G and X should be finite.
Theorem VIII.7 (Stabilizer–Orbit Theorem). If G acts transitively on X and
both are finite, then
|G| = | orbG (x)| · | StabG (x)|
for every point x of X. If the action is transitive, so that there is only one orbit
X, this becomes
|G| = |X| · | StabG (x).
I won’t prove this formally, but give some ideas.
If x, y ∈ X are in the same orbit, then StabG (x) and StabG (y) are conjugate
subgroups as you show in homework. So in particular, they have the same size. This
explains why in the theorem it is not important which x one takes: all stabilizers
are conjugate to one another.
Next you cluster the elements of G in such a way that g, g 0 are in the same
cluster if and only if gx = g 0 x. Note that g · StabG (x) all end up in the same cluster
since they all send x to gx. Note that these sets are just the left cosets relative to
H := StabG (x). One then checks (easy but detailed) that g, g 0 belonging to different
H-costets rules out the possibility of gx = g 0 x. So, the clusters all are of the same
74 VIII. WEEK 8: GROUP ACTIONS
size as StabG (x). So, G is partitioned into clusters of size | StabG (x)|, and elements
g, g 0 in different clusters produce different output gx 6= g 0 x when multiplied against
x. But the collection of all outputs Gx is just the orbit of x. So, as the theorem
claims, |G| = | orbG (x)| · | StabG (x)|.
(This all should remind you much of Lagrange’s Theorem and its proof. In
fact, this proof here is the proof for Lagrange’s Theorem if you take X to be the
coset space for the subgroup H as in the example above. Then the Stabilizer-Orbit
Theorem becomes Langrange’s: ”|G| = |G/H| · |H|”. In reverse, this theorem and
its proof is simply Lagrange applied to G and its subgroup H := StabG (x)).
Finally, if G acts transitively, there is only one orbit, and so orbG (x) = X.
Now we discuss fixed point counts. Recall that for g ∈ G, the set FixX (g) is
the points of X that are unmoved by g, so gx = x. Let us alo write X/G for the
orbit space of X under G. This is just the set of all orbits, the notation sugesting
that X/G arises from X by clustering elements of X where clusters are orbits. The
following theorem addresses the question of counting the number of orbits.
Theorem VIII.9 (Burnside). If G acts on X and both are finite, then the size
of the orbit space is
1 X
|X/G| = | FixX (g)|.
|G|
g∈G
P Again, I won’t give a very formal proof but the main ideas. Let us count
g∈G | FixX (g)| as follows. Look at the collection of pairs (g, x) in the Cartesian
product G × X for which gx = x. Let F be the collection of all such pairs. We can
sort them by the separate g, or the separate x. If we sort them by P g then we get
clusters FixX (g) and so the number of all such pairs is precisely g∈G | FixX (g)|.
But if we cluster by x, then each cluster has the form {g ∈ G|gxP = x} and that is
exactly StabG (x). So, if we now sum this over all x we get x∈X | StabG (x)|. Of
course, these two counts must agree:
X X
| StabG (x)| = | FixX (g)|.
x∈X g∈G
We now need to interpret the sum on the left a bit differently. From the
Stabilizer–Orbit Theorem, if we let G just act on the orbit Gx of x, we know
P |G| = | StabG (x)| · | orb
that PG (x)|. So, restricting the sumP to the orbit of x, we get
x∈orbG (x) | StabG (x)| = x∈orbG (x) |G|/| orbG (x)| = |G| x∈orbG (x) 1/| orbG (x)| =
|G|. P
So, orbit by orbit, the expression x∈X | StabG (x)| contributes one copy of |G|.
If you sum over all orbits, this is |G| times the sum P of the number of orbits. The
latter is |X/G|, and so we find that |G| · |X/G| = x∈X | StabG (x)|. Combined
1
P
with the equation |X/G| = |G| g∈G | FixX (g)| this shows the Burnside Theorem.
Note that there is very little “power” in this proof, it relies on 2 ways of counting
the same thing.
VIII. WEEK 8: GROUP ACTIONS 75
Example VIII.10. How many different dice can one make with labels 1 through
6 on them? It turns out, this question is made for Mr Burnside.
First off, if the die can’t move, there are 720 = 6! ways to paint numbers on the
faces of the cube. The problem is that dice can move, and so many of the seemingly
different dice will turn out to be the same.
Let us write X for the 720 different dice that we painted. Let G be the sym-
metry group of the cube, it moves the dice around and has |G| = 24 elements.
If 2 dice are truly differently labeled, they would not look the same under any
symmetry. So they would not be in the same G-orbit. In other words, we want to
count the size of the orbit space X/G.
If we plan to use the Burnside Theorem, we need to study the fixed points of
all motions. Note that a “fixed point” is now a labeling of the cube that looks the
same no matter what we do with that cube. But it is clear that every rigid motion
of the cube will move a face and in fact several. So there are no g witha fixed point.
Unless, of course, you took g to be the identity motion, which has every labeling
as a fixed point. So, in the Burnside formula there is exactly one summand that
contributes anything, namely the one that belongs to g = e. And the summand for
g = e is | FixX (e)| = |X| = 720. All other summands belong to a g without fixed
1
points and contribute 0. So the formula says |X/G| = 24 (720 + 0 + . . . + 0) = 30.
The example makes clear a special case of the Burnside Theorem:
Corollary VIII.11. If X acts on G and no element e 6= g ∈ G has any fixed
point, then |X/G| = |X|/|G|.
Review
• Week 1
– induction, well ordering
– modular arithmetic
– primes and irreducibles in a domain
– Euclidean algorithm in Z, gcd, lcm, relative prime (coprime)
• Week 2
– symmetries of an object and composition of symmetries
– group (axioms), and Cayley table
– cancellation property in groups
– exampls: symmetry groups, KV4 , GL(n, R), vector spaces, (Z/nZ, +),
U (n), Cn , free groups, Zn
– Abelian groups, cyclic groups
– order of a group, and of elements in a group
– subgroup
– product group G × H
– Aut(G), the relabelings of G that preserve the Cayley table, a group
with composition
• Week 3
– the Euler φ-function
– the number of elements in the cyclic group Cn that have order d
(distinction for d|n and d 6 |n)
– the number of subgroups of a given size in Cn
– the number of generators for Cn
– φ(Z/pqZ) = φ(Z/pZ) · φ(Z/qZ) if gcd(p, q) = 1
– if a = a1 · ak then Ca1 × Ca2 × · · · Cak = Ca provided that the ai are
pairwise coprime
– solving x mod n = a mod n, a mod m = b mod m when m, n co-
prime
– U (mn) = U (m) × U (n) if coprime
– |U (pk )| = pk−1 (p − 1), and why
– φ : Z/mnZ → (Z/mZ) × (Z/nZ) via 1 + mnZ 7→ (1 + mZ, 1 + nZ) is
isomorphism provided m, n coprime
• Week 4
– left and right cosets of G relative to the subgroup H; coset space
G/H
– morphisms ψ : G → G0 and a list of examples
– know how to test whether multiplication by k ∈ N gives a morphism
Z/mZ → Z/nZ
– conjugation by a, g 7→ aga−1
77
78 REVIEW
This begins the second part of the course, where we study structures that
allow both addition and multiplication. The standard example is Z, with Q, R, C
following closely behind.
Definition IX.1. A ring is a set R with a binary operation + : R × R → R
called addition and a second binary operation · : R × R → R called multiplication
such that
(1) (R, +) is an Abelian group;
(2) multiplication is associative, (r · s) · t = r · (s · t) for all r, s, t ∈ R;
(3) the distributivity law is intact: r(s+t) = r ·s+r ·t and (r +s)·t = r ·t+s·t
for all r, s, t ∈ R;
(4) there is a neutral element for multiplication, written 1R , with 1R · r = r =
r · 1R for all r ∈ R.
It is perhaps useful to make some comments here.
• We denote 0R (or just 0) the neutral element for addition in R, and write
(−a) for the additive inverse of a ∈ R. Note the following two facts.
a · 0 = a · (0 + 0) = (a · 0) + (a · 0), so a · 0 = 0;
0 = a · 0 = a · (1 + (−1)) = a · 1 + a · (−1),
so that (−1) · a is the additive inverse of a. We usually denote it by −a
and write b − a for b + (−1) · a.
• We will almost exclusively look at commutative rings, which are those
where r · s = s · r for all r, s ∈ R. But there is a general consensus that
non-commutative rings are important enough for not being disqualified
from the start.
• Some people no not require the existence of 1R . Rings without multiplica-
tive identity are not difficult to find, but they lack key features of rings
that we want to discuss in our remaining chapters.
• One thing to note is something that a ring need not have, and that is
multiplicative inverses. We are not saying that invereses must not exist
(after all, 1R is always its own inverse!); we just concede that they may
not exist in all cases. Do not confuse + and ·; + is always commutative
by definition.
• However, if an element a ∈ R does have a multiplicative inverse, this
inverse is unique, because if a0 , a” both are inverses to a ∈ R then a0 =
a0 · e = a0 · (a · a”) = (a0 · a) · a” = e · a” = a”.
Example IX.2. Here is a list of standard rings that come up all the time in
mathematics. The first three are all commutative.
79
80 IX. WEEK 9: INTRODUCTION TO RINGS
• The rings after which all others are modelled is (Z, +, ·), the set of integers
with usual addition and multiplication.
• The three collections Q, R, C of rational, real and complex numbers re-
spectively are all rings as well. They are rather special rings, since in
contract to Z, every non-zero number in these three rings does have a
multiplicative inverse (whereas in Z that is only the case for ±1.)
• The groups Z/nZ are also all rings, with addition and multiplication of
cosets.
• A collection of non-commutative matrix rings arises for each number n
and choice of coefficient ring K as follows. Let Mn (K) be the set of all n×n
matrices with entries in K. Then usual matrix addition and multiplication
has the usual properties, which are those listed in Definition IX.1 above.
Note that in general A · B 6= B · A so that Mn (K) is not commutative.
• Another collection of rings are the polynomial rings K[x1 , . . . , xn ] over
a chosen coefficients field K. These are commutative rings, and their
elements are the polynomials in the variables x1 , . . . , xn which have coef-
ficients in K.
• A type of ring we will not look at much is popular in analysis: the set of
all real-valued functions on the interval [0, 1]. Addition and multiplication
is pointwise, which means that (f + g)(x) is declared as f (x) + g(x) and
likewise for multiplication.
√
Example IX.3. Here is an example of an extension ring. Let Z[ −1] be√the set
√ numbers that have both real and imaginary value integer. So, Z[ −1] =
of complex
{a + b −1 with a, b ∈ Z}. You might want to√ think of this as a “vector space of
dimension
√ 2 over Z, √ spanned by 1 ∈ Z and √−1”. So, we add componentwise:
(a + b −1) + (c + d −1) = (a + c) + (b + d) −1. Multplication has a bit of a
surprise, as it does
√ not go componentwise,
√ but instead like√for complex numbers in
general: (a + b −1) · (c + d −1) = (ac − bd) + (bc + ad) −1. This is the ring of
Gaussian integers.
Definition IX.4. If in a ring R we have a, b ∈ R both nonzero, but ab = 0,
then we call a and b zero-divisors.
Most of the rings listed in examples here do not have zero-divisors. The excep-
tions are: Z/nZ if n is not prime; Mn (R) in the case n > 1 and also in the case
that R itself has zero-divisors; the polynomial ring R[x1 , . . . , xn ] in the case that R
has zero-divisors. You might want to check these three claims explicitly by finding
one example of zerodivision in each of the three scenarios.
Definition IX.5. A commutative ring that has no zero-divisors is called a
domain.
Note that if a ∈ R has an inverse, then a cannot be a zero-divisor. Indeed, if
ab = 1 and ca = 0 then c = c · 1 = cab = 0 · b = 0.
Definition IX.6. Consider 1R , 1R +1R , 1R +1R +1R , . . .. This sequence might
or might not contain the element 0R . If it does, there is a smallest number c ∈ N+
such that adding 1R c times gives 0R . We call this c the characteristic of R.
If this sequence never produces 0R we say that the characteristic of R is zero.
Lemma IX.7. If R is a domain, its characteristic is a prime number or zero.
IX. WEEK 9: INTRODUCTION TO RINGS 81
Definition IX.8. A commutative ring that has multiplicative inverses for each
nonzero element is called a field.
Proof. We need to show that the absence of zero-divisors forces the presence
of inverses when R is finite. Take a ∈ R nonzero. then multiplication by a gives
a permutation of the elements of R. Indeed, let r1 , . . . , rt be the complete list
of nonzero elements of R. Then ar1 , . . . , art is another list of elements of R. No
expression ari can be zero, sine a 6= 0 and ri 6= 0, and R is supposed to be a domain.
Also, there is no repetition on this list since if ari = arj then a(ri − rj ) = 0 and
a 6= 0 now forces (ri − rj ) = 0 as otherwise we would be looking at zero-divisors
which can’t exist in a domain. So, the second list is a permutatin of the first list,
because both list all nonzero elements of R. (This is where finiteness is used: if
R were infinite we could not argue like this. For example, the multiples of 2 are
not a permutation of the nonzero integers. But in a finite set, if you list as many
different elements of S as the set has, you listed them all). It follows, that one of
the elements on the second list is 1R , which amounts to saying that there is ri ∈ R
with a · ri = 1R .
Remark IX.10. A postscript of this proof goes like this: let R have p elements.
Then the nonzero elements are a group with multiplication, since the theorem
assures the esitence of inverses. This group has p − 1 elements. So Lagrange says
that if a is a nonzero element of R then its mutiplicative order divides p − 1. In
particular, there is a power c such that ac = 1R . But then ac−1 · a = 1R and so the
inverse of a is actually a power of a.
Example IX.11. In the same way the Gaussian integers √ are an extension of
the integers, one can mak√ extensions of fields. For example, Q[ 2] is the collection
√
of all √ √ a, b ∈ Q. One adds componentwise,
expressions a + b 2 with √ (a + b√2) +
(c + d 2) = (a + c) + (b√+ d) 2 and multiplies according to (a + b 2) · (c + d 2) =
(ac + 2bd) + (bc + ad) 2. √
√ √
Note how one computes inverses here: (a + b 2)−1 = (a+(a−b √
2)
√
2)(a− 2)
= a−b 2
a2 −2b2 =
a b
√
a2 −2b2 + a2 +2b2 2 is of the required form. Recall that we proved that there cannot
be rational numbers a, b with a2 = 2b2 and so the numerator is nonzero.
82 IX. WEEK 9: INTRODUCTION TO RINGS
Example
√ IX.12.
√ One can do this also with modular numbers. Let R =
(Z/3Z)[ 2]. Here, 2 stands for a symbol whose square is the coset of 2. (Note
that there is no element in Z/3Z whose square is the coset of 2, just like there was
no rational number whose square was 2.) √
This is a ring with 9 elements, the possible expressions of the form a + b 2
with a, b ∈ Z/3Z. You calculate exactly as expected, always going modulo 3. √
√ √
So for example, the inverse of 2 + 1 2 is 2+1 2 = (2+1 2)(2−1 2) = 2−1
1√ 2−1 2 √
√
4−2
2
=
√ √ √ √ √
1 − 2 2 = 1 + 1 2. And indeed, (2 + 1 2)(1 + 1 2) = 1 + 0 2.
CHAPTER X
Recall that we insist that our rings have a (multiplicative) 1. (All rings have a
neutral element for + (which we write as 0), since R with + is a group).
Definition X.1. A ring morphism is a function f : R → R0 from one ring to
another such that it is a morphism of groups (R, +) → (R0 , +), and moreover it
respects ring multiplication: f (r1 r2 ) = f (r1 )f (r2 ).
Examples of such things abound.
√
• the inclusions Z ,→ Q ,→ R ,→ CC and the inclusions Z ,→ Z[ −1] ,→ C;
• the surjection Z → Z/nZ sending k to k + nZ for any n;
• if m|n, the surjection Z/nZ → Z/mZ sending k + nZ to k + mZ;
• complex conjugation;√ √ √ √
• the “conjugation” Z[ 2] → Z[ 2] sending a + b 2 to a − 2 and any
similar constructs;
• the polynomial map C[x, y] → C[t] that sends x 7→ t2 , y 7→ t3 ;
• If O is the collection of real functions defined on the real line, then any
a ∈ R induces an evaluation morphism a : O → R that sends f (x) ∈ O
to the value f (a) of f at a.
Example X.2. Recall that there are rings of positive characteristic. If char(R) =
p > 0 is prime, there is the Frobenius morphism Frob : R → R that sends r ∈ R to
Frob(r) = rp . That this is then a morphism is due to freshman’s dream in algebra:
(x + y)p = xp + y p in characteristic p, since by the binomial theorem every missing
term of (x + y)p is a multiple of p.
Definition X.3. If f : R → R0 is a ring morphism, its kernel is the elements
of R that are sent to 0 ∈ R0 by f .
Definition X.4. An ideal in a ring R is a subset I ⊆ R such that
• I is a subgroup of R with respect to addition;
• For all x ∈ I and all r ∈ R, the product xr is in I.
Remark X.5. A standard way of producing ideals is as follows. Let f1 , . . . , fk
be elements of the ring R. Then let I be the set of all R-linear combinations you
can make from f1 , . . . , fk . In other words, I is made precisely of all things like
r1 f1 + . . . + rk fk where r1 , . . . , rk run through all possible elements in R. Then I
is an ideal: sums of such things as well as differences of such things are such things
again, and multiplying any such element by an arbitrary ring element gives another
thing of this type.
It is important to note that it is allowed for an ideal to have infinitely generators.
Often, one can simplify such a situation to finitely many generators, but not always.
The rings we consider all will have only ideals that are finitely generated, but
proving this can be dicey (although we will prove it in some nice cases).
83
84 X. WEEK 10: IDEALS AND MORPHISMS
For example, the multiples of 641 are an ideal of Z. So are the C[x, y]-linear
combinations of x3 − y 2 and x3 + y 4 in C[x, y].
Proposition X.6. The kernel of a ring morphism is an ideal.
Proof. That the kernel of a ring morphism f : R → R0 is a subgroup of
R follows straight from the fact that f is a group morphism. Now take x ∈ I
and r ∈ R. Then f (x) = 0 and so f (x)f (r) = 0 and so f (rx) = 0 and so
rx ∈ ker(f ) = I.
We next turn this around and use ideals to make factor rings and morphisms.
Definition X.7. Let I ⊆ R be an ideal. The factor ring R/I is the group R/I
together with multiplication (x + I)(y + i) = xy + I. There is an induced morphism
π : R → R/I that sends r ∈ R to r + I.
That this construction indeed produces a ring is not difficult to see. One
basically needs to check that multiplication is well-defined (this means that if x+I =
x0 + I and y + I = y 0 + I then xy + I = x0 y 0 + I, but that is quite easy.
If f : R → R0 is a ring morphism and I an ideal of R and J an ideal of R0 , then
inspection shows that
• f (I) may not be an ideal in R0 (for example, 2Z is an ideal in Z but when
you inject Z ,→ R then the even integers are no longer an ideal; make sure
you believe this, it is due to the fact that products of integers and reals
are often not integer).
• the preimage f −1 (J) in contrast is always an ideal of R. This is seen
as follows. Since f is a group morphism, the preimage is a group. Take
x ∈ f −1 (J) and y ∈ R. Then f (xy) = f (x)f (y) ∈ J · R0 = J and so
xy ∈ f −1 (J) as required.
• If char(R) = n then there is a natural morphism Z/nZ ,→ R induced by
sending 1 ∈ Z to 1 ∈ RR and using the morphism rule.
The main structure theorem for ideals says:
Theorem X.8. If I is an ideal of R then under the natual surjection π : R →
R/I the ideals of R/I correspond to the ideals of R that contain I. More precisely,
if J is an ideal of R that contains I then the quotient group J/I is an ideal of R/I.
In reverse, if J/I is an ideal of R/I then the preimage f −1 (J/I) is an ideal of R.
For example, if R = Z and I = 6Z then R/I has 4 ideals: the whole ring R0 =
Z/6Z, the zero idewal {0+6Z} and two interesting ideals J2 = {0+6Z, 2+6Z, 4+6Z}
and J3 = {0 + 6Z, 3 + 6Z}. To J2 corresponds the ideal 2Z of R, and it indeed
contains I. To J3 corresponds the ideal 3Z of R, and indeed it contains I. The
only ideals that contain I are I, 2Z, 3Z, Z. The first of these corresponds to the
zero idealin R/I and the last one to the whole of R/I.
We come now to talk about certain special types of ideals.
Definition X.9. A prime ideal of a ring R is an ideal P such that if a, b ∈ R
with ab ∈ P then at least one of a, b is in P .
Being a prime ideal is equivalent to saying that R/P is a domain. (There are
a, b ∈ R but not in P such that ab ∈ P if and only if in R/P we have (a+P )(b+P ) =
0 + P which can happen if and only if R/P is not a domain).
X. WEEK 10: IDEALS AND MORPHISMS 85
Proof. The second claim, as mentioned previously, follows directly from the
structure theorem of factor rings. The proof for the first claim is analogous to the
proof of the preceding lemma.Namely, if I is a prime ideal and (a + I)(b + I) = 0 + I
in R/I then we must have ab ∈ I and so by primeness of I one of a, b is in I, and
thus one of a + I, b + I is zero in R/I. If I is not prime, there are a, b ∈ R that are
not in I but with ab ∈ I. Then (a + i)(b + I) = 0 + I are zerodivisors and so R/I
is not a domain.
Theorem X.16. Every ideal in Z and in Z/nZ is generated by one element.
Proof. Suppose the ideal I ⊆ Z contains a and b. By the Euclidean algorithm,
it also contains their gcd g. On the other hand, a, b are multiples of g and so we
see that any ideal that contains a, b also contains gcd(a, b) and conversely.
Iterating this argument, ha, b, ci = hgcd(a, b), ci = hgcd(a, b, c)i, and ha, b, c, di =
hgcd(a, b), c, di = hgcd(a, b, c), di = hgcd(a, b, c, d)i, and in this way every finite gen-
erator set a1 , . . . , ak for an ideal can be replaced by the single generator given by
the gcd of all ai .
Now imagine an infinite list a1 , a2 , . . . , an , . . .. We know that
gcd(a1 ) ≥ gcd(a1 , a2 ) ≥ gcd(a1 , a2 , a3 ) . . . ≥ 0.
It follows that this sequence of ≥ symbols reaches a point (say, when the index is
k) from where onwards each ≥ is actually a =.
What this means is that gcd(a1 , . . . , ak ) divides ak+1 , ak+2 , . . .. But then
ak+1 , ak+2 , . . . are already in the ideal generated by a1 , . . . , ak and we can say
that
ha1 , a2 , . . . , an , . . .i = ha1 , ldots, ak i = hgcd(a1 , . . . , ak )i
is a generated by one element.
Definition X.17. Ideals generated by one element are called principal. The
theorem says that Z has only principal ideals. Since Z is a domain (has no zerodi-
visors), it is referred to as a principal ideal domain.
Remark X.18. You will prove in homework that ideals in Z/nZ are also all
principal.
CHAPTER XI
ri0
if and only if we have ri = for all i. Note that for large i this is automatic since
eventually all coefficientsPare zero.
∞
Given a polynomial i=0 ri xi , there is a largest index d for which rd is nonzero,
and this index d we call the degree deg(f ) of thePpolynomial. If d = deg(f ) then
∞
we usually write r0 + r1 x + . . . + rd xd instead of i=0 ri xi , and call rd the leading
coefficient lc(f ) of f (x). .
We add polynomials degree by degree:
∞
X ∞
X ∞
X
ri xi + ri0 xi = (ri + ri0 )xi .
i=0 i=0 i=0
1. Euclidean rings
Definition XI.3. A domain has a Euclidean measure if there is a function
δ : R \ {0} → N that satisfies
(1) δ(a) ≤ δ(ab) for all a 6= 0 6= b in R;
87
88 XI. WEEK 11, EUCLIDEAN RINGS
Example XI.8. Let’s do one that is a bit more thrilling. Let’s compute gcd of
x10 − 1 and of x6 − 1 in Q[x].
geerated by 2 and x, I = {2a + xb|a, b ∈ R}. Let is assume for the moment that I
is principal, generated by the polynomial f . So that means that 2 is a multiple of
f and also x is a multiple of f ,
2 = f g, x = f h,
with g, h ∈ R. Plugging x = 0 into the second equation, 0 = f (0)h(0) and so one of
f (0) and h(0) has to be zero. Plugging x = 0 into the first equation, 2 = f (0)g(0)
and the says that f (0) is not zero, hence h(0) = 0. But then h is a multiple of x,
h = xk with k ∈ R. Together then, x = f h = f kx says that 1 = f k when dividing
out x. That says that the ideal hf i of multiples of f contains 1. Since we labor
under the belief that h2, xi = hf i, 1 should be a linear combination of 2 and x,
1 = 2a + xb with a, b ∈ R. Then evaluation at x = 0 gives 1 = 2a(0) which is not
possible since a(0) is an integer.
It follows that h2, xi is not principal and so R cannot have a Euclidean algorithm
and thus cannot have a Euclidean measure.
Definition XI.14. If R is a domain, then its ring of fractions is the ring whose
elements are fractions of the form f /g with f, g ∈ R but g nonzero. Addition and
muliplication are exactly as you would think.
So, for example, the ring of fractions of the domain Z is the ring of rational
numbers, and the ring of fractions of the polynomial ring R[x] is the ring of rational
functions with real coefficients.
Let us note that in a ring of fractions, f /g has inverse g/f unless f = 0. This
means that a ring of fractions of a domain is actually a field. Note also that there is
an inclusion of rings of R into its ring of fractions that sends f ∈ R to the fraction
f /1R . This is the natural generalization of the inclusion Z ,→ Q via z 7→ z/1.
The notion of a ring of fractions comes up in the proof of the next result.
Theorem XI.15 (The Gauß Lemma). If R is a domain and has unique fac-
torization, then so does R[x].
Proof. The idea is as follows. Let K be the ring of fractions of R. Then we
have an inclusion R[x] ,→ K[x] that is a ring morphism. Given f (x) a polynomial in
R[x], we can now also read it as a polynomial in K[x]. But as K is a field, we have
shown that K[x] is Euclidean, and therefore a UFD. So, in K[x] we can uniquely
factor f (x) = g1 (x) · · · gk (x) where each gi (x) is a polynomial in x with coefficients
in K, and no gi (x) can be factored further in K[x].
The question is how to translate this back into R[x]. The problems are: first
off, no gi (x) might be in R[x] (because of the fractions in the coefficients); secondly,
if we ever manage to make a translation, why is the resulting factorization for f (x)
in R[x] unique?
Skipping all of the details, the main part of the work consists now in showing
that one can rearrange the denominators in the various gi (x) such that after the
rewriting all factors have coefficients in R. In other words, if a product of poly-
nomials with fractional coefficients only has “whole” coefficients, then one rewrite
to a factorization with whole coefficients in each factor. For example, we can take
x1 − in Z[x] and rewrite in Q[x] as (2x + 1)(x/2 − 1/2), but by moving around the
1/2 we can also rewrite to x2 − 1 = (x + 1)(x − 1).
The official statement to be proved is:
Lemma XI.16. If f ∈ R[x] can be factored as f (x) = g(x)h(x) with g, h ∈ R[x],
then any prime element p ∈ R that divides f coefficient by coefficient, must divide
one of g or h coefficient by coefficient.
The proof of the lemma proceeds by an iterated induction on the degrees of
f, g, h.
With the lemma in hand one can prove that a factorization of f in K[x] always
yields a related factorization in R[x]. Uniqueness is then rather easy. You might
look at the proof of the Gauß Lemma in anyt textbook, if you are curious.
CHAPTER XII
1. Divisibility
Let R be a commutative ring with 1, and take an element f (x) in the polynomial
ring R[x]. For ever r ∈ R there is an evaluation morphsim
εr : R[x] → R
that sends f (x) to the element f (r) in R. It is immediately clear that if a polynomial
f (x) is a multiple of x − r then εr (f ) = 0 simply because εr (x − r) = 0. So, the
kernel of εr contains at least all multiples of x − r (that is, the ideal generated by
x − r).
It turns out that this kernel is precisely the ideal generated by x − r. The
argument is the following. Write f (x) = a0 + a1 x + . . . + ad xd , d the degree of f ,
and suppose εr (f ) = 0. Since εf (x − r) = 0 as well, then for arbitrary g(x) ∈ R[x]
we also have εr (f (x) − g(x) · (x − r)) = 0, since we can do the plug-in process
separately in the two polynomials.
Let’s pick a g1 (x) in such a way that f1 (x) := f (x) − g1 (x) · (x − r) has degree
less than d. By construction, εr (f ) = εr (f1 ). Now repeat: find g2 (x) such that
f2 (x) := f1 (x) − g2 (x) · (x − r) has degree less than deg(f1 ). Keep going. At the
end of the day, this must stop, because when you found a fk (x) that is constant,
you can’t keep the iteration going.
We have εr (f ) = εr (f1 ) = εr (f2 ) = . . . = εr (fk ) and that fk (x) is a constant.
But as a constant, pluggin in has no effect. So εr (f ) = fk . This says that the
remainder that you get when you divide f (x) by x − r (that is hat fk (x) really is!)
is precisely the value of f (x) at input x = r.
Lemma XII.1. Let f (x) ∈ R[x] and choose r ∈ R. The value f (r) is the
remainder of division of f (x) by x − r.
Going back to the kernel of our morphism εr , this lemma says that: f (x) ∈
ker(εr ) happens if and only if f (x) has remainder zero when dividing by x − r. But
the latter statement is just a euphemism for “f (x) is a multiple of x − r. So,
ker(εr ) = R[x] · (x − r).
Definition XII.2. If f (x) ∈ ker(εr ) we call r a root of f (x) in R.
Roots can be funny.
Example XII.3. (1) The roots of x2 − 1 in Z/12Z are 1 + 12Z.5 + 12Z, 7 +
12Z, 11 + 12Z. So a degree 2 polynomial can have more than 2 roots. The culprit
is the fact that Z/12Z is not a domain. Note that this is also reflected in possible
factorizations: x2 − 1 = (x − 1)(x + 1) = (x − 5)(x − 7) in Z/12Z.
(2) (x + 1)2 has only one root, -1, but with multiplicity two.
93
94 XII. DIVISIBILITY, FIELD EXTENSIONS
In a while, we will go and try to manufacture new fields from old. As a stepping
stone we need to know when a polynomial is irredecble. We will be mainly concerned
with R = Z and Z/pZ with prime p ∈ Z. Note that over R = Z/pZ we can
actually go and test all elements of the field on whether they are roots as there are
finitly many things to test. Over Q that is much harder. Here is a basic test for
irreducibility.
Lemma XII.5. Let f (x) ∈ Z[x] be given, and assume that the coefficients of
f (x) have no common factor. If you can find a prime number p ∈ Z such that
f (x) mod p is irreducible and of the same degree as f , then f is irreducible in Z[x]
and even in Q[x]..
Proof. The Gauß Lemma says that if we can show that f (x) is irreducible in
Z[x] then it is also irreducible in Q[x]. So we focus on irreducibility on Z[]x.
Suppose f = gh with g, h ∈ Z[x]. Then take this equation and reduce modulo
p to get f (x) mod p = (g(x) mod p)(h(x) mod p). Since f (x) mod p is supposed
to be irreducible, this new equation must have one of g mod p, h mod p be a unit.
But units must have degree zero. So between g(x) mod p and h(x) mod p, one
has degree zero. However, that means that the other factor must have degree
deg(f mod p) = deg(f ), and so one of g or h themselves has degree deg(f ). That
now means that the other one of g, h has degree zero and so is an integer.
We have shows that f (x) can only be factored in Z[x] as (integer) times (poly-
nomial of degree deg(f )). Since the coefficients of f have no common factor by
hypothesis, the integer factor is a unit and we are done.
Remark XII.6. The lemma says that irreducibility “lifts” from Z/pZ[x] to
Z[x]. It is not true that reducibility also lifts. Many polynomials are reducible
modulo p but irreducible over Z. There are even polynomials in Z[x] that are
irreducible but become reducible modulo every prime p. You will work through
one such example (f (x) = x4 + 1) in the homeworks.
1. DIVISIBILITY 95
Proof. You learned in discrete math that number cp,k is the number of of
ways to pick k things from p given ones, and that there is an explicit formula
p!
cp,k = k!(p−k)! . In particular, the numerator is a multiple of p. It then suffices to
show that the denominator is not a multiple of p (since p is prime!). To see this,
note that k and p − k are both less than k, and so neither k! nor (p − k)! has a
factor divisible by p. Since p is prime, and p divides neither factor, it also does not
divide the product.
g, h ∈ F[x] are such that g and h in R are zerodivisors in R. That means that
gh ∈ hf i, so that gh = αf for some α ∈ F[x]. But then gh is divided by f and since
f is prime, f muts divide one of them, say f |g. But then g = 0 in R, so R has no
zerodivisors and is a domain.
Now take g ∈ F[x] with g nonzero in R and look for an inverse. Since g
is nozero, f can’t divide g. Since F[x] is a Euclidean ring, the gcd of f, g is a
linear combination of f, g. But this gcd is 1 since f is irreducble. Thus there are
polynomails a(x), b(x) with f (x)a(x) + g(x)b(x) = 1. Read this module hf i to get
g · b = 1. So b is an inverse to g.
From the definition, it is clear that Kron(F, f ) has a basis given by 1, x, . . . , xdeg f −1 .
Now we prove the last claim. Let us make a ring morphism from F[x] to F(β)
by sending x to β and any element of F to itself. The kernel is the polynomials p(x)
for which p(β) = 0. These are the multiples of the minimal polynomial of β. This
minimal polynomial is a divisor of f (x) and since f (x) is irreducible, this minimal
polynomial is f (s) itself.
It follows that we can actually make a ring morphism from Kron(F, f ) to F(β)
by sending the coset of x to β and all elements of F to themselves.
As we know, Kron(F, f ) is a field, and so is by definition F(β). So we have
one field contained in another, which makes the bigger field a vector space over the
smaller one. The bigger one is generated over F by 1, β, β 2 . . . , β deg f −1 , and each
of these is in the image of the morphism, xi 7→ β i . So, the morphism is a surjective
inclusion, hence an isomorphism.
which allows to rewrite all powers of α in terms of just second, first, and zeroth
powers. So, R is as a vector space over Q spanned by 1, α, α2 .
Supposedly, f has a root in R called α. Let;s find the other factor: divide y 3 −2
by y − α. We find a quotient of y 2 + αy + α2 .
It is entirely reasonable to ask whether this quadric splits further. In other
words, does f have a further root in R? To find out, pick an element of R; it will
look like aα2 + bα + c with rational numbers a, b, c. Now take this and plug it for
y into y 2 + αy + α2 . After sifting through the mess, you find that you obtained
(using that α3 = 2 of course)
α2 (b2 + 2ac + b + 1) + α(2a2 + 2bc + c) + 1(c2 + 4ab + 2a).
We are asking whether this can be zero for suitable choices of a, b, c ∈ Q. This is
here not totally easy, and in general (other Kronecker extensions) can be extremely
hard.
Here we can argue as follows: if the displayed expression is zero, then the three
expressions b2 +2ac+b+1, 2a2 +2bc+c = 2a2 +c(2b+1), c2 +4ab+2a = c2 +2a(2b+1)
are all zero. But then 2a2 = −c(2b+1) and c2 = −2a(2b+1) gives c/2a = 2a2 /c2 . So
(2a/c)3 = 2 and as weknow there are no rational numbers a, c suchthat (2a/c)3 = 2.
It follows that the quadric does not have any further roots in R.
CHAPTER XIII
field for x3 − 2 we need first β and then additionally a Kronecker extension that
catches a root of x2 + βx + β 2 .
It is time for the following concept.
Definition XIII.5. If F ⊆ E is an extension of fields, E is a vector space over
F. We denote the vector space dimension of E over F by [E : F] and call it the
degree of the extension.
Clearly, the degree of a Kronecker extension Kron(F, f ) is the degree of the
polynomial f , since Kron(F, f ) has the basis 1, x, x2 , . . . , xdeg f −1 .
Example XIII.6. Kron(Q, x3 − 2) is degree 3 over Q; Split(Q, x3 − 2) is degree
6 over Q; Split(F, f ) is of degree at most (deg(f ))! over F.
Example XIII.7. It is definitely possible for a cubic polynomial to split in a
degree 3 extension (within the first Kronecker extension of the iterative splitting
process).
Let F = Z/2Z and choose f = x3 + x + 1. Since f has no roots in F (check
that!), it has no linear factors over Z/2Z. Since it is degree 3, it has no factors and
is thus irreducible.
Let β be the Kronecker root for f in K := Kron(F, f ). Then β 2 and β 2 + β are
also roots of f inside K. This can be seen by stupidly plugging in: (β 2 )3 +(β 2 )1 +1 =
β 6 + β 2 + 1 = (β 3 + β + 1)2 since we are in characteristic 2. But β 3 + β + 1 = 0.
Similarly, (β 2 + β)3 + (β 2 + β)1 + 1 = β 6 + 3β 5 + 3β 4 + β 3 + β 2 + β + 1 =
(β + β 2 + 1) + (β 5 + β 3 + β 2 ) + (β 4 + β 2 + β 1 ) (remember that 2=0 here!). Each
6
On the third root in the example: if a cubic f (x) has 2 roots r1 , r2 in some
field, the third root is also in the field, and you can find it by longly dividing f (x)
first by x − r1 and then what you got by x − r2 . You’ll be left with x − r3 , and that
is what I did.
It is kind of clear that if a polynomial has a factor of the sort (x − r)k , then r
is a root and of multiplicity at least k. So, before one investigates multiplicity, one
should perhaps split the polynomial as far as one can into irreducibles.
As one sees by examples, it is ofetn interesting to take a polynomial over one
ring R and ask for its roots in a bigger ring. For eample, we know that we need to
look inside C for roots of x2 + 1.
Some strange things can happen.
Example XIII.9. Let K = Z/pZ. Then xp −1 (which over the complex numbers
has the p different roots of 1 as solutions) is equal to (x−1)p (because of freshman’s
dream in characteristic p). So, it has only one root, x = 1, and that with multiplicity
p.
Stranger yet, there are polynomials that are irreducible and yet have multiple
roots in a suitable larger ring.
Example XIII.10. Let K = Z/pZ(t). So, p = 0 in our ring, t is a variable, and
we are looking at the rational functions in t with coefficients in Z/pZ. (Recall that
“rational function” means “quotient of two polynomials).
Now look at the polynomial xp − t. In K, this has no roots because the p-th
root of a variable is not expressible as a quotient of 2 polynomials in that variable.
We will show in√a bit, that xp − t is also irreducible. Let’s make the field bigger,
say K̃ = Z/pZ( p t) the rational functions with Z/pZ coefficients in the symbol p-th
root of t. √ √ √
If p = 2, we have (x − 2 t)(x − 2 t) = x2 − 2x 2 t + t = x2 + t = x2 − t since
2 = 0. √ √ √2
For p = 3 we have√(x − 3 t)3 = x3 − 3x3 3 t + 3x 3 t − t = x3 − t since 3 = 0.
In general, (x − p t)p = xp + p(stuff) − t where the middle part is the stuff
that comes from√ the binomial theorem for i = 1. . . . , p − 1. In all cases then,
xp − t = (x − p t)p has a p-fold root in K̃ while it was irreducible over K.
Here is a way for testing whether a polynomial can ever have multiple roots.
The prime in the theorem denotes taking the derivative according to the rules of
calculus: product rule and power rule. (You might ask “What other rules might I
possible want to use for a derivativ, isn’t that a stupid thing to say?”. You are sort
of right. There are no other rules one should ever use. But the fact is that in some
environments, calculus seems like a dubious activity to engage in. For example, in
Z/3Z[x], what could “differentiation” mean? Normally, a derivative is a limit, but
in Z/pZ there are only finitely many “numbers”, so limits are very limited in their
nature. . . )
Theorem XIII.11. Let F be a field and f (x) ∈ F[x]. Then f (x) has a double
root in some (perhaps mysterious) extension field E ⊇ F if and only if gcd(f, f 0 ) is
not 1.
In other words, if f, f 0 are coprime then f has single roots in any field.
Proof. If in some extension field E we have (x − r)2 |f (x) (so r ∈ E is a
multiple root) write f (x) = (x − r)2 · g(x). Then taking derivatives, we have
f 0 (x) = 2(x − r)g(x) + (x − r)2 g(x) = (x − r)[2g(x) + (x − r) · g 0 (x)] is a multiple
of x − r. Of course, so is f itself, and so x − r divides both f, f 0 and hence must
divide their gcd. This means that a multiple root in an extension field prevents the
gcd of f, f 0 being 1.
102 XIII. SPLITTING FIELDS AND EXTENSION TOWERS
Now suppose the gcd of f, f 0 is not 1, or in other words, some g(x) of positive
degree divides both f and f 0 . Then let E be the Kronecker extension on F for
any irreducible factor of g(x). In E, g(x) has the Kronecker root β, and so g(x)
is a multiple of x − β and also f (x) is a multiple of x − β. So we can write
f (x) = h(x)(x − β). Then the derivative of f is f 0 (x) = h0 (x)(x − β) + h(x). Now
plug in x 7→ β. We know that (x − β)|g(x)|f 0 (x), so f 0 (β) = 0. But then h(β) must
also be zero. That says that (x − β) divides h(x), and so f (x) = (x − β)h(x) has
x − β twice as factor. So β is a double root of f in E.
Remark XIII.12. In characteristic zero (when K contains Q) an irreducible
polynomial is relatively prime to its own derivative, because the derivative is a
nonzero polynomial of lower degree, and so cannot have a common divisor with the
irreducible f .
In prime characteristic, the derivative f 0 (x) can be zero without f being a
constant. For example, the polynomial xp −t from the example above has derivative
zero, since (xp )0 = pxp−1 and p = 0. (Note that we take x-derivatives, so (t)0 = 0
as t and x do not relate in that example!) In that case, then, we have gcd(f mf 0 ) =
gcd(f, 0) = f .
Definition XIII.13. A polynomial f (x) with coefficients in the field F is sep-
arable if f does not have multiple roots in any extension field of F. Any other
polynomial is inseparable.
The choice of “separable” indicates that separable polynomials have their roots
“separated out” in any extension: the roots never equal one another. In character-
istic zero, “irreducible” implies “separable”. But in characteristic p, separability is
an actual condition. It is a fact that over a finite field, “irreducible” still implies
“separable”, but in infinite fields of characteristic p one needs to be careful.
CHAPTER XIV
1. Minimal Polynomials
Recall that a field extension F ⊆ E makes E a vector space over F. (Think of
R ⊆ C). The start of our investigations is based on
Definition XIV.1. If F ⊆ E is a field extension it is called algebraic if for any
α ∈ E the powers 1, α, α2 , . . . of α are linearly dependent over F.
There are many field extensions that are not algebraic. For example, Q ⊆ R
is not finite, because the powers 1, π, π 2 , . . . of π have no Q-linear dependence.
Another way is to say that π does not occur as a root of a polynomial in Q[x], π is
transcendental. Finiteness of a field extension indicates that the two fields are not
too far from one another, in a sense to be discussed this week and next.
For now we note the obvious
Theorem XIV.2. If F ⊆ E is algebraic, then for any α ∈ E there is a monic
irreducible polynomial f (x) ∈ F[x] such that f (α) = 0.
Note that previously we started with a polynomial and looked for roots; this is
now the other way round.
Pk
Proof. If the powers of α are dependent, there is an expression i=0 ri αi
Pk
that equates to zero. The polynomial i=0 ri xi is what we are looking for. It can
be made monic (have lead coefficient 1) by dividing out the actual lead coefficient.
Note that this division does not affect the vanishing of the polynomial at x = α.
If we have several polynomials that vanish for x = α, then their gcd has the
same property. So, the gcd of all polynomials that vanish at x = α is one such as
well, and clearly the one of lowest degree. (Note: if I is the ideal of all polyno-
mials vanishing at x = r then the generator for this ideal—principal since F[x] is
Euclidean—is the one we want.)
Definition XIV.3. If F ⊆ E is a field extension, and if the powers of α ∈ F are
linearly dependent over F then the monic polynomial f (x) ∈ F[x] of minimal degree
with f (α) = 0 is the mimimal polynomial of α over F and denoted minpolF (α).
√ 2
Note that√ √ to know F in this definition: minpolR ( −1) = x +1,
one really needs
but minpolC ( −1) = x − −1. We note for future purposes that the second
(complex) minimum polynomial divides the first (real).
Definition XIV.4. If F ⊆ E is a field extension and if the powers of α ∈
E are linearly dependent over F then α is algebraic over F. Elsewise we call α
transcendental over F.
103
104 XIV. WEEK 14: MINIMAL POLYNOMIALS AND FINITE FIELDS
2. Finite Fields
Example XIV.9. Let F = Z/2Z and take f (x) = x3 + x + 1, g(x) = x3 + x2 + 1.
It is easy to check that neither f nor g have a root in F, and so (as cubics) are
irreducible.
Kron(F, f ) = F[x]/hf i has 8 = 23 elements {0, 1, x, x + 1, x2 , x2 + 1, x2 + x, x2 +
x+1}. The same is true for Kron(F, g) = F[y]hgi, but we must not confuse elements
in the different extensions because in the first, we go modulo f and in the other we
go modulo g. I intentionally write different variables x, y here.
Let α be a the Kroenecker root for f , so α = x mod hf i. Write β for the
Kronecker root of g, so β = x mod hgi.
So α is a root of f , who else? f (x) : (x − α) = x2 + αx + (α2 + 1), which we
call f2 (x). Then if you plug α2 into f2 (x), you get zero, so α2 is a root of f2 and
also then of f (x). The remaining root can be found as α2 + α. (As a test, if you
multiply out (x − α)(x − α2 )(x − α2 − α) you get f (x) back, using that f (α) = 0.)
So Kron(F, f ) is actually the splitting field of f over F.
Now plug x − 1 into f (x). You get (x − 1)3 + (x − 1) + 1 = x3 + x2 + 1 = g(x).
So, g has roots equal to thos of f shifted up by 1. They are α + 1, α2 + 1, α2 + α + 1.
In particular, Kron(F, f ) is also the splitting field for g(x). So there is no real
difference between α+1 in Kron(F, f ) and β ∈ Kron(F, g). There is only one degree
3 extension of F.
Every element of GF(p, e) is equal to its pe -th power, and so every element has
e
a p -th root in GF(p, e).
One has GF(p, 1) = Z/pZ.
The degree of the field extension [GF(p, e) : GF(p, 1)] is e. In consequence,
GF(p, e) is the Kronecker extension Kron(F, g(x)) for every irreducible polynomial
of degree e.
e
Sketch. Existence Take any splitting field K for f (x) := xp − x (for
example, a suitable field inside an iteration of Kronecker extensions). Denote this
set by GF(p, e).
Note that if a, b are both roots of f (x) then that is also true for a±b and ab and
a/b provided that b 6= 0. (Why? For ± the binomial theorem gives you p-divisible
e
coefficients in (a ± b)p in all but the first and last term. For ab and a/b this is very
easy.) It follows that GF(p, e) is closed under + and −, and under multiplication
and division. So, this set of roots is a field (This
√ is really weird and only happens
over finite fields. For example, the field Q[ 2] has lots and lots of elements that
are not roots of x2 − 2. . . )
Splitting By construction, f has all its roots in GF(p, e), so GF(p, e) is the
smallest field over which f splits. It follows that GF(p, e) is the splitting field.
e
Size The gcd of f (x) and f 0 (x) = pe xp −1 − 1 is 1, since pe = 0 and so f 0 (x) =
−1. So, f has no multiple roots in any extension, and in particular not in GF(p, e).
So, GF(p, e) is full of single roots of f (x) and so must have pe elements (it is a
plitting field!).
e
Uniqueness Any field with pe elements has ap − a = 0 by the argument of the
e
remark above, so any field with p elements is the splitting field of f .
(1) Since f (a) = 0 for all a ∈ GF(p, e), each element agrees with its own pe -th
power.
(2) If e = 1, we want the splitting field over Z/pZ of xp − x. But Little Fermat
says that ap = a for each a ∈ Z/pZ. So all roots of xp − x are in Z/pZ and we need
no extension.
(3) A vector space with pe elements over a field of p elements has to have
e basis vectors. Let g be an irreducible polynomial of degree e over Z/pZ. Its
Kronecker extension is a field extension of order e, so has pe elements, and so must
e
be a splitting field for xp − x. So Kron(F, g) = GF(p, e)
Corollary XIV.12. The nonzero elements U (p, e) of GF(p, e) form an Abelian
group with respect to multiplications. This group is cyclic.
Proof. The first sentence is clear since fields have commutative multiplication
and every nonzero element in a field has an inverse.
As Abelian group, U (p, e) can be writen as Z/a1 Z×· · ·×Z/ak Z with a1 |a2 | · · · |ak ,
by FTFGAG. If q is an element of this product group, it has order ak . So the el-
ements of U (p, e) have their ak -th power equal to the identity. That means, they
are roots of xak − 1. So elle elements of GF(p, e) are roots to xak +1 − x. But such
a polynomial can have only ak + 1 roots, and we know GF(p, e) is the set of these
roots, pe in number. So pe = ak + 1. So ak = pe − 1. But a1 · a2 · · · ak shouldbe
pe − 1 = |U (p, e)|, and that means that k = 1 and so U (p, e) is cyclic.
Example XIV.13. Let p = 2 and take f (x) = x4 + x + 1. Since f (0) = f (1) =
1 ∈ Z/2Z, f has no linear factors.
2. FINITE FIELDS 107
Example XIV.17. Let’s look at the finite fields inside the field of size 224 .
They are the fields of sizes 212 , 28 , 26 , 24 , 23 , 22 , 21 . The containment relations are
GF(2, 1) ⊆ GF(2, 2) ⊆ GF(2, 4) ⊆ GF(2, 8) ⊆ GF(2, 24),
GF(2, 1) ⊆ GF(2, 3) ⊆ GF(2, 6) ⊆ GF(2, 12) ⊆ GF(2, 24),
and additional containments GF(2, 2) ⊆ GF(2, 6) and GF(2, 4) ⊆ GF(2, 12).
CHAPTER XV
Galois
1. The Frobenius
In a ring of characteristic p > 0 (such as in GF(p, e) or indeed a ring containing
Z/pZ), we have
(a + b)p = ap + pp
since the intermediate terms arising from the binomial theorem all are multiples of
p, hence zero. It follows that
Frob : GF(p, e) → GF(p, e),
γ 7→ γ p
is a morphism of additive groups. Since clearly 1p = 1 and (γγ 0 )p = γ p (γ 0 )p , the
Frobenius also respects the multiplicative structure. It is therefore a ring moor-
phism.
Theorem XV.1. The p-Frobenius (p-th power map) is a field isomorphism
Frob : GF(p, e) → GF(p, e)
for any e.
If e0 |e and GF(p, e0 ) therefore sits inside GF(p, e) then the Frobenius sends
elements of this subfield into the subfield.
The e-fold iteration of the Frobenius on GF(p, e) is the identity map. One can
interpret this as
the group Z/eZ acts on GF(p, e) by sending the coset of t to the
t-fold iteration of Frob.
The elements of GF(p, e) that are fixed by Frob are exactly the elements of
GF(p, 1) = Z/pZ. More generally, the elements of GF(p, e) that are fixed under the
k-fold iteration of the p-th power map are precisely the elements of GF(p, gcd(e, k)).
Suppose α ∈ GF(p, e) is the root of a polynomial f (x) with coefficients in Z/pZ.
Then αp is a root of f (x) as well. In fact, iterating the p-th power map will produce
all other roots in GF(p, e) of f (x). In other words,
The orbit of α ∈ GF(p, e) under the action of Z/eZ above is the
set of all roots of the minimal polynomial of α over Z/pZ.
If k ∈ N one can define an action of Z/eZ via λ(t, α) = (αp )t . The orbits under
this action are the roots of the minimal polynomial of α over GF(p, gcd(e, k)).
Proof. Let α be in the kernel of Frob. Then αp = 0 and since a field has no
zerodivisors, α = 0. So Frob is injective. By the isomorphism theorem, the image
of Frob is isomorphic to GF(p, e). But that means it has pe elements, and hence fills
out the target field. So, Frob is bijective and hence an isomorphism. It permutes
the elements of GF(p, e).
109
110 XV. GALOIS
e
All elements in GF(p, e) satisfy αp = α. If e0 |e, the elements of GF(p, e0 ) inside
e0
GF(p, e) are characterized by being those elements for which αp = α already. Take
e 0 e0 e0
such α and raise it to the p-th power. Then note that (αp )p = αp·p = (αp )p =
αp . In other words, Frob(α) belongs to GF(p, e0 ) again. So the isomorphisms that
the Frobenius induces on the various fields GF(p, −) are compatible with inclusions.
e
The e-fold iteration of Frob sends α ∈ GF(p, e) to αp = α, so it is the identity
on GF(p, e). It follows that we can read Frob as a group actiion of Z/eZ on the
t
elements of GF(p, e) via λ(t mod eZ, α) 7→ Frobt (α), which is nothing but αp .
Since Frobe is the identity on GF(p, e), Frobk acts the same way as Frobgcd(e,k) .
(You should make sure you believe this before going on. It can be seen via the
Euclidean algorithm: Frobk (α) = Frobk−e (α); now iterate). If αp = α then α
is a root of xp − x, and there are exactly p of those, the elements of Z/pZ =
gcd(e,k
GF(p, 1). If Frobk (α) = α then α is a root of xp − x and therefore belongs to
GF(p, gcd(k, e)).
Suppose e0 |e, so GF(p, e0 ) sits inside GF(p, e). If f (α) = 0 and the coefficients
0
of f come from a field GF(p, e0 ) then the coefficients ci satisfy Frobe (ci ) = ci .
e0
ci αi produces under e0 -fold Frobenius that 0 = ci (αp )i .
P P
Thus, 0 = f (α) =
0
In other words, Frobe (α) is a root to the same polynomial as α. Since the degree of
α over GF(p, 1) is the product of the degree of α over GF(p, e0 ) with e/e0 , it follows
that the degree of the minimal polynomial of α over GF(p, e0 ) is e/e0 . This implies
0
that iterating Frobe on α makes it circle throuugh all the roots of f . (If it did not
move through all roots, one could take the roots it moves thorugh and construct a
minimal polynomial of lower degree, which cannot be).
• elements in GF(3, 4) that are not in GF(3, 2): these come in quadruplets
as the roots of the 18 irreducible quartics.
Let us take the irreducible quadric x2 + 1, and let α be the Kronecker root
of GF(3, 4) = Kron(Z/3Z, x4 − x3 − 1) for f (x) = x4 − x3 − 1. In other words,
α = x. Let’s try to find a copy of GF(3, 2) inside this field. This would require, for
example, finding the roots to x2 + 1 (one of the three irreducible quadrics above).
We calculate
(α3 + α2 + 1)2 + 1 = α6 + 2α5 + α4 + 2α3 + 2α2 + 1 + 1
= α2 · (α3 + 1) + 2α5 + α4 + 2α3 + 2α2 + 2
= 3α5 + α4 + 2α3 + 3α2 + 2
= α4 − α3 − 1 = 0.
It follows that α3 + α2 + 1 is a root of x2 + 1. (The other root is 2(α3 + α2 + 1).)
So, inside GF(3, 4) the copy of GF(3, 2) consists of the Z/3Z-linear combinations
of 1 and β := α3 + α2 + 1. These are the 9 elements
0, 1, 2, β, β + 1, β + 2, 2β, 2β + 1, 2β + 2.
Now look at what the Frobenius (third power map) does to them:
03 = 0,
3
1 = 1,
23 = 8 = 2,
(α3 + α2 + 1)3 = α9 + α6 + 1 = . . . = 2α3 + 2α2 + 2,
((α3 + α2 + 1) + 1)3 = . . . = 2α3 + 2α2 + 0,
((α3 + α2 + 1) + 2)3 = . . . = 2α3 + 2α2 + 1,
3 2 3
(2(α + α + 1)) = . . . = α3 + α2 + 1,
(2(α3 + α2 + 1) + 1)3 = . . . = α3 + α2 + 2,
(2(α3 + α2 + 1) + 2)3 = . . . = α3 + α2 + 0
So, the third-power map flips them about in pairs. The 3 pairs correspond to the
roots of the 3 irreducible quadrics above.
Now let us look what the Frobenius does to general elements of GF(3, 4), those
that do not live in smaller fields. As a starter, we look at what happens to α itself
under iterates of Frob. It is clear that Frob(α) = α3 and we leave it like that since
we can’t rewrite polynomials of degree less than four.
Then Frob(Frob(α)) = α9 and that can be rewritten (with labor) as α3 +α2 +2α.
The third power of this is α3 + 2α2 + 1, and the Frobenius sends this last guy to
α. So the Frobenius action circles
α 7→ α3 7→ α3 + α2 + 2α 7→ α3 + 2α2 + 1 7→ α.
These 4 elements are the roots of x4 + x3 + 1, since we took one such root, and
applied Frobenius. (Frobenius takes the equation x4 − x3 − 1 = 0 and turns it into
(x3 )4 − (x3 )3 − 1 = 0, so that if you “Frobenius a root” then you get a root back).
The same sort of thing happens to the roots of the other 17 irreducible quadrics:
the Frobenius circles them within their lucky clover leaf, preserving that they are
roots to whatever quadric they are roots of.
So, Z/4Z (the 4 is because e = 4 and the 4-th power of the Frobenius is the
identity) acts on the 81 elements of GF(3, 4). Three elements are fixed points, there
112 XV. GALOIS
are thre orbits of size 2 (pairing the roots of the quadrics) and there are 18 orbits
of size 4 (the 18 quadruplets that occur as roots of the irreducible quartics)
2. Symmetries
3. Applications
Stuff for later
The coset of 0 in Z/nZ is “the zero” in this new system of numbers, since adding
it to any coset does not change the coset. As seen in Example XV.3 above, it is
possible that this new zero shows up as a product of nonzero inputs, a phenomenon
not encountered in the integers.
Definition XV.5. If a, b are in Z/nZ, with neither a nor b divisible by n, then
they are called zerodivisors if ab = 0.
113
114 STUFF FOR LATER
This ability to multiply to zero in Z/nZ of course comes from the fact that we
equate (every multiple of) n with zero. So, a composite n will allow for products
to be zero (that is, multiples of n) in several ways. We try to understand by way
of an example.
Example XV.6. Let n = 6; then 2 · 3 = 0.
Indeed, in order to prepare what is to come in a bit, let’s list all multiples of 2:
2 · 0 = 0, 2 · 1 = 2, 2 · 2 = 4, 2 · 3 = 0, 2 · 4 = 2, 2 · 5 = 4.
The reason that 2 was capable to yield 0 when multiplied with a nonzero coset
was of course that 2 has an interesting common factor with 6. In the general case,
suppose a is a coset in Z/nZ and we look for another element b ∈ Z/nZ such that
ab = 0. If we set gcd(a, n) = d and if d happens to be greater than 1, then we can
write n = d · e and so a · e is a multiple of d · e = (de) = n = 0. But a multiple of 0
must be 0 itself.
On the other hand, pick now an a such that gcd(a, n) = 1. This means by
Proposition I.18 that there are integers α, β with aα + nβ = 1. Reading this
“modulo n”, we get a · α + (nβ) = 1. Naturally, (nβ) = n · β = 0. So, a · α = 1. It
follows that for any b ∈ Z, a · (bα) = b. This says that every single coset in Z/nZ
is the result of some coset being multiplied by a.
Let’s try to understand what this means. There are n cosets in Z/nZ, each
of which you can multiply with a. The process of multiplication produces all n
of these (provided gcd(a, n) = 1). It follows there is exactly one coset that when
multiplied by a gives you any given coset b. In particular, there is only one coset
that when multiplied gives 0 (and of course this one coset is 0 itself).
Putting it all together, we have proved most of the following theorem:
Theorem XV.7. If gcd(a, n) = 1 then multiplication by a is a bijection on
Z/nZ. In other words, for each b ∈ Z/nZ there is exactly one x ∈ Z/nZ such that
a · x = b. Yet in other words, a becomes a unit in Z/nZ.
Conversely, if gcd(a, n) = d > 1 then multiplication by a is neither surjective
nor injective. There are exactly n/d different cosets that arise through multiplication
by a, and each is d times output of such a multiplication. In this case, a is a
zerodivisor in Z/nZ.
Example XV.8. Let n = 6. The numbers a that have gcd(a, n) = 1 are living
in the cosets 1 and 5. Everyone is a multiple of 1 for obvious reasons, and everyone
is also a multiple of 5 because 5 = −1.
The multiples of 2 are {0, 2, 4}, and these are also exactly the multiples of 4.
Note that each one of {0, 2, 4} is a multiple of both 2 and 4 in 2 = gcd(6, 2) =
gcd(6, 4) ways. For example, 4 = 4 × 1 = 4 × 4 and also 4 = 2 × 2 = 2 × 5.
The multiples of 3 are 3 and 0, and each of {0, 3} arises 3 = gcd(3, 6) times as
multiple. For example, 3 = 3 × 1 = 3 × 3 = 3 × 5.
Exercise XV.9. For n = 10 and a = 1, 2, . . . , 9 determine
(1) which cosets in Z/10Z are multiples of a;
(2) how many cosets in Z/10Z are multiples of a and express these numbers in
terms of a and 10.
STUFF FOR LATER 115
Theorem XV.7 implies that the units of Z/nZ are exactly the cosets of those
numbers between 1 and n − 1 inclusive that are relatively prime to n. All other
cosets exhibit ambiguity (at best) or impossibility (at worst) when trying to divide
by them. Which case happens depends on the two cosets to be divided. For
example, in Z/4Z, trying to divide by 2 one fails when the input is 1 or 3 while one
gets too many suggestions when one divides 2 by 2 (namely, 1 and 3) or when one
divides 0 by 2 (namely, 0 and 2).
In order to explain this behavior, we shall need the following observation:
Exercise XV.10. Prove that lcm(a, b) · gcd(a, b) = ab.
Now suppose a · x = b has at least one solution, so ax − b is a multiple of
n. If you added c = n/ gcd(a, n) to x then we calculate: a(x + c) = ax + ac =
b + (an/ gcd(a, n)) = b + lcm(a, n) = b since lcm(a, n) = 0 (like any other multiple
of n) represents the coset of zero. It follows that besides x all expressions x + i · c
are also solutions to ax = b.
How many such are there? On the face of it, infinitely many but recall that
x+i·c and x+j·c are in the same coset of Z/nZ as soon as (x+i·c)−(x+j·c) = (i−j)c
is a multiple of n. That of course happens exactly if i − j is a multiple of n/c. So,
there are n/c different cosets x, x + c, . . . , x + ((n/c) − 1)c that all solve ax = b.
(Of course, n/c = gcd(a, n) by definition of c).
Exercise XV.11. Group the elements of Z/24Z in such a way that two cosets
a, b are in the same group exactly when their sets of multiples {1a, 2a, 3a, . . .} and
{1b, 2b, 3b, . . .} agree as sets (perhaps after reordering). Describe in words each
group.
Remark XV.12. The Euclidean algorithm can also be carried out in the poly-
nomial ring R[x]; the idea of size (absolute value for integers) in the Archimedean
principle is then taken over by the degree of the polynomial. The relevant statement
is then:
For all polynomials a(x), b(x) in R[x] there are q(x), r(x) ∈ R[x]
such that a(x) = b(x)q(x)+r(x) and 0 ≤ deg(r) ≤ deg(b)−1.
The polynomials q(x) and r(x) are furnished by the method of (polynomial) long di-
vision. Exactly as for integers, one can work this division process into an algorithm
to compute the gcd between polynomials.
Exercise XV.13. Compute the gcd between
(1) x3 + 1 and x1 + 1;
(2) x3 + 1 and x2 + 1;
(3) x3 + 1 and x4 + 1;
(4) x3 + 1 and x5 + 1;
(5) x3 + 1 and x6 + 1;
(6) x3 + 1 and xn + 1 for any natural number n (this will require to consider
cases depending on the remainder of division of n by 6.
3.2. Cartesian Products, Euler’s φ-function, Chinese Remainder. We
wish to find a formula for the number of cosets in Z/nZ that are units. By Theorem
XV.7, we need to count the numbers on the list 1, . . . , n − 1 that are coprime to n.
For this, recall the Euler φ-function from Definition I.32.
116 STUFF FOR LATER
with gcd(a, b) = gcd(b, c) = gcd(c, a) = 1, one has two options. Either, take the
souped-up version of the Chinese Remainder Theorem which we state below. Or,
one first solves x = r mod a and x = s mod b as above and then y = x mod ab and
y = t mod c again as above.
Here is the multiverse formulation of the Chinese Remainder Theorem; its proof
is in parallel to its little brother above.
Theorem XV.19 (Chinese Remainder Theorem). Let n1 , n2 , . . . , nt be pairwise
coprime numbers. Choose values a1 , . . . , at . Then the set of integers x which leave
remainder ai when divided by ni for all i are the elements in the coset x + n1 · · · nt Z
determined as follows. Let N = n1 · · · nt and set Ni = N/ni . Find, for each i, a
solution xi to the equation Ni · xi = 1 mod ni . Then
t
X
x= xi Ni ai .
i=1