Datasets:
Tasks:
Text Generation
Modalities:
Text
Sub-tasks:
language-modeling
Languages:
English
Size:
100K - 1M
License:
\chapter{Foundations} | |
\label{foundations} | |
The present foundational chapter will introduce the notion of a ring and, | |
next, that of a module over a ring. These notions will be the focus of the | |
present book. Most of the chapter will be definitions. | |
We begin with a few historical remarks. Fermat's last theorem states that the | |
equation | |
\[ \label{ft} x^n + y^n = z^n \] | |
has no nontrivial solutions in the integers, for $n \ge 3$. We could try to | |
prove this by factoring the expression on the left hand side. We can write | |
\[ (x+y)(x+ \zeta y) (x+ \zeta^2y) \dots (x+ \zeta^{n-1}y) = z^n, \] | |
where $\zeta$ is a primitive $n$th root of unity. Unfortunately, the factors | |
lie in $\mathbb{Z}[\zeta]$, not the integers $\mathbb{Z}$. Though | |
$\mathbb{Z}[\zeta]$ is still a \emph{ring} where we have notions of primes and | |
factorization, just as in $\mathbb{Z}$, we will see that prime factorization | |
is not always unique in $\mathbb{Z}[\zeta]$. (If it were always unique, then we | |
could at least one important case of Fermat's last theorem rather easily; see | |
the introductory chapter of \cite{Wa97} for an argument.) | |
For instance, consider the ring | |
$\mathbb{Z}[\sqrt{-5}]$ of complex numbers of the form $a + b\sqrt{-5}$, where | |
$a, b \in \mathbb{Z}$. Then we have the two factorizations | |
\[ 6 = 2 \cdot 3 = (1 + \sqrt{-5})(1 - \sqrt{-5}). \] | |
Both of these are factorizations of 6 into irreducible factors, but they | |
are fundamentally different. | |
In part, commutative algebra grew out of the need to understand this failure | |
of unique factorization more generally. We shall have more to say on | |
factorization in the future, but here we just focus on the formalism. | |
The basic definition for studying this problem is that of a \emph{ring}, which we now | |
introduce. | |
\section{Commutative rings and their ideals} | |
\subsection{Rings} | |
We shall mostly just work with commutative rings in this book, and consequently | |
will just say ``ring'' for one such. | |
\begin{definition} | |
A \textbf{commutative ring} is a set $R$ with an addition map | |
$+ : R \times R \to R$ and a multiplication map $\times : R \times R \to R$ | |
that satisfy the following conditions. | |
\begin{enumerate} | |
\item $R$ is a group under addition. | |
\item The multiplication map is commutative and distributes over addition. | |
This means that $x \times (y+z) = x \times y + x\times z$ and $x \times y = y | |
\times x$. | |
\item There is a \textbf{unit} (or \textbf{identity element}), denoted by | |
$1$, such that $1 \times x = x$ for all $x \in R$. | |
\end{enumerate} | |
We shall typically write $xy$ for $x \times y$. | |
Given a ring, a \textbf{subring} is a subset that contains the identity | |
element and is closed under addition and multiplication. | |
\end{definition} | |
A \emph{noncommutative} (i.e. not necessarily commutative) ring is one | |
satisfying the above conditions, except possibly for the commutativity | |
requirement $xy = yx$. For instance, there is a noncommutative ring of | |
$2$-by-$2$ matrices over $\mathbb{C}$. We shall not work too much with noncommutative rings in | |
the sequel, though many of the basic results (e.g. on modules) do generalize. | |
\begin{example} | |
$\mathbb{Z}$ is the simplest example of a ring. | |
\end{example} | |
\begin{exercise}\label{polynomial} Let $R$ be a commutative ring. | |
Show that the set of polynomials in one variable over $R$ is a commutative | |
ring $R[x]$. Give a rigorous definition of this. | |
\end{exercise} | |
\begin{example} | |
For any ring $R$, we can consider the polynomial ring $R[x_1, \ldots, x_n]$ | |
which consists of the polynomials in $n$ variables with coefficients in $R$. | |
This can be defined inductively as $(R[x_1, \dots, x_{n-1}])[x_n]$, where the | |
procedure of adjoining a single variable comes from the previous | |
\cref{polynomial}. | |
\end{example} | |
We shall see a more general form of this procedure in \cref{groupring}. | |
\begin{exercise} | |
If $R$ is a commutative ring, recall that an \textbf{invertible element} (or, somewhat | |
confusingly, a \textbf{unit}) $u \in R$ is an element such | |
that there exists $v \in R$ with $uv = 1$. Prove that $v$ is necessarily | |
unique. | |
\end{exercise} | |
\begin{exercise} \label{ringoffns} | |
Let $X$ be a set and $R$ a ring. The set $R^X$ of functions $f:X \to R$ is a | |
ring. \end{exercise} | |
\subsection{The category of rings} | |
The class of rings forms a category. Its morphisms are called ring homomorphisms. | |
\begin{definition} | |
A \textbf{ring homomorphism} between two rings $R$ and $S$ as a map | |
$f : R \to S$ that respects addition and multiplication. That is, | |
\begin{enumerate} | |
\item $f(1_R) = 1_S$, where $1_R$ and $1_S$ are the respective identity | |
elements. | |
\item $f(a + b) = f(a) + f(b)$ for $a, b \in R$. | |
\item $f(ab) = f(a)f(b)$ for $a, b \in R$. | |
\end{enumerate} | |
There is thus a \emph{category} $\mathbf{Ring}$ whose objects are commutative | |
rings and whose morphisms are ring-homomorphisms. | |
\end{definition} | |
The philosophy of Grothendieck, as expounded in his EGA \cite{EGA}, is that one should | |
always do things in a relative context. This means that instead of working | |
with objects, one should work with \emph{morphisms} of objects. Motivated by | |
this, we introduce: | |
\begin{definition} | |
Given a ring $A$, an \textbf{$A$-algebra} is a ring $R$ together with a | |
morphism of rings (a \textbf{structure morphism}) $A \to R$. There is a category of $A$-algebras, where a | |
morphism between $A$-algebras is a ring-homomorphism that is required to commute with the structure | |
morphisms. | |
\end{definition} | |
So if $R$ is an $A$-algebra, then $R$ is not only a ring, but there is a way | |
to multiply elements of $R$ by elements of $A$ (namely, to multiply $a \in A$ | |
with $r \in R$, take the image of $a $ in $R$, and multiply that by $r$). | |
For instance, any ring is an algebra over any subring. | |
We can think of an $A$-algebra as an arrow $A \to R$, and a morphism from $A | |
\to R$ to $A \to S$ as a commutative diagram | |
\[ \xymatrix{ | |
R \ar[rr] & & S \\ | |
& \ar[lu] A \ar[ru] | |
}\] | |
This is a special case of the \emph{undercategory} construction. | |
If $B$ is an $A$-algebra and $C$ a $B$-algebra, then $C$ is an $A$-algebra in a | |
natural way. Namely, by assumption we are given morphisms of rings $A \to B$ | |
and $B \to C$, so composing them gives the structure morphism $A \to C$ of $C$ | |
as an $A$-algebra. | |
\begin{example} | |
Every ring is a $\mathbb{Z}$-algebra in a natural and unique way. There is a | |
unique map (of rings) $\mathbb{Z} \to R$ for any ring $R$ because a | |
ring-homomorphism is required to preserve the identity. | |
In fact, $\mathbb{Z}$ is the \emph{initial object} in the category of rings: | |
this is a restatement of the preceding discussion. | |
\end{example} | |
\begin{example} | |
If $R$ is a ring, the polynomial ring $R[x]$ is an $R$-algebra in a natural | |
manner. Each element of $R$ is naturally viewed as a ``constant polynomial.'' | |
\end{example} | |
\begin{example} | |
$\mathbb{C}$ is an $\mathbb{R}$-algebra. | |
\end{example} | |
Here is an example that generalizes the case of the polynomial ring. | |
\begin{example} | |
\label{groupring} | |
If $R$ is a ring and $G$ a commutative monoid,\footnote{That is, there is a | |
commutative multiplication on $G$ with an identity element, but not | |
necessarily with inverses.} then the set | |
$R[G]$ of formal finite sums $\sum r_i g_i$ with $r_i \in R, g_i \in G$ is a | |
commutative ring, called the \textbf{moniod ring} or \textbf{group ring} when | |
$G$ is a group. | |
Alternatively, we can think of elements of $R[G]$ as infinite sums $\sum_{g \in | |
G} r_g g$ with $R$-coefficients, such that almost all the $r_g$ are zero. | |
We can define the multiplication law such that | |
\[ \left(\sum r_g g\right)\left( \sum s_g g\right) = | |
\sum_h \left( \sum_{g g' = h} r_g s_{g'} \right) h. | |
\] | |
This process is called \emph{convolution.} We can think of the multiplication | |
law as extended the group multiplication law (because the product of the | |
ring-elements corresponding to $g, g'$ is the ring element corresponding to | |
$gg' \in G$). | |
The case of $G = | |
\mathbb{Z}_{\geq 0}$ is the polynomial ring. | |
In some cases, we can extend this notion to formal infinite sums, as in the | |
case of the formal power series ring; see \cref{powerseriesring} below. | |
\end{example} | |
\begin{exercise} | |
\label{integersinitial} | |
The ring $\mathbb{Z}$ is an \emph{initial object} in the category of rings. | |
That is, for any ring $R$, there is a \emph{unique} morphism of rings | |
$\mathbb{Z} \to R$. We discussed this briefly earlier; show more generally that | |
$A$ is the initial object in the category of $A$-algebras for any ring $A$. | |
\end{exercise} | |
\begin{exercise} | |
The ring where $0=1$ (the \textbf{zero ring}) is a \emph{final object} in the category of rings. That | |
is, every ring admits a unique map to the zero ring. | |
\end{exercise} | |
\begin{exercise} | |
\label{corepresentable} | |
Let $\mathcal{C}$ be a category and $F: \mathcal{C} \to \mathbf{Sets}$ a | |
covariant functor. Recall that $F$ is said to be \textbf{corepresentable} if | |
$F$ is naturally isomorphic to $X \to \hom_{\mathcal{C}}(U, X)$ for some | |
object $U \in \mathcal{C}$. For instance, the functor sending everything to a | |
one-point set is corepresentable if and only if $\mathcal{C}$ admits an | |
initial object. | |
Prove that the functor $\mathbf{Rings} \to \mathbf{Sets}$ assigning to each ring its underlying set is | |
representable. (Hint: use a suitable polynomial ring.) | |
\end{exercise} | |
The category of rings is both complete and cocomplete. To show this in full | |
will take more work, but we can here describe what | |
certain cases (including all limits) look like. | |
As we saw in \cref{corepresentable}, the forgetful functor $\mathbf{Rings} \to | |
\mathbf{Sets}$ is corepresentable. Thus, if we want to look for limits in the | |
category of rings, here is the approach we should follow: we should take the | |
limit first of the underlying sets, and then place a ring structure on it in | |
some natural way. | |
\begin{example}[Products] | |
The \textbf{product} of two rings $R_1, R_2$ is the set-theoretic product $R_1 | |
\times R_2$ with the multiplication law $(r_1, r_2)(s_1, s_2) = (r_1 s_1, r_2 | |
s_2)$. It is easy to see that this is a product in the category of rings. More | |
generally, we can easily define the product of any collection of rings. | |
\end{example} | |
To describe the coproduct is more difficult: this will be given by the | |
\emph{tensor product} to be developed in the sequel. | |
\begin{example}[Equalizers] | |
Let $f, g: R \rightrightarrows S$ be two ring-homomorphisms. Then we can | |
construct the \textbf{equalizer} of $f,g$ as the subring of $R$ consisting of | |
elements $x \in R$ such that $f(x) = g(x)$. This is clearly a subring, and one | |
sees quickly that it is the equalizer in the category of rings. | |
\end{example} | |
As a result, we find: | |
\begin{proposition} | |
$\mathbf{Rings}$ is complete. | |
\end{proposition} | |
As we said, we will not yet show that $\mathbf{Rings}$ is cocomplete. But we | |
can describe filtered colimits. In fact, filtered colimits will be constructed | |
just as in the set-theoretic fashion. That is, the forgetful functor | |
$\mathbf{Rings} \to \mathbf{Sets}$ commutes with \emph{filtered} colimits | |
(though not with general colimits). | |
\begin{example}[Filtered colimits] | |
Let $I$ be a filtering category, $F: I \to \mathbf{Rings}$ a functor. We can | |
construct $\varinjlim_I F$ as follows. An object is an element $(x,i)$ for $i | |
\in I$ and $x \in F(i)$, modulo equivalence; we say that $(x, i)$ and $(y, j)$ | |
are equivalent if there is a $k \in I$ with maps $i \to k, j \to k$ sending | |
$x,y$ to the same thing in the ring $F(k)$. | |
To multiply $(x, i)$ and $(y,j)$, we find | |
some $k \in I$ receiving maps from $i, j$, and replace $x,y$ with elements of | |
$F(k)$. Then we multiply those two in $F(k)$. One easily sees that this is a | |
well-defined multiplication law that induces a ring structure, and that what we | |
have described is in fact the filtered colimit. | |
\end{example} | |
\subsection{Ideals} | |
An \emph{ideal} in a ring is analogous to a normal subgroup of a | |
group. As we shall see, one may quotient by ideals just as one quotients by | |
normal subgroups. | |
The idea is that one wishes to have a suitable \emph{equivalence relation} on a | |
ring $R$ such that the relevant maps (addition and multiplication) factor | |
through this equivalence relation. It is easy to check that any such relation | |
arises via an ideal. | |
\begin{definition} | |
Let $R$ be a ring. An \textbf{ideal} in $R$ is a subset $I \subset R$ that | |
satisfies the following. | |
\begin{enumerate} | |
\item $0 \in I$. | |
\item If $x, y \in I$, then $x + y \in I$. | |
\item If $x \in I$ and $y \in R$, then $xy \in I$. | |
\end{enumerate} | |
\end{definition} | |
There is a simple way of obtaining ideals, which we now describe. | |
Given elements $x_1, \ldots, x_n \in R$, we denote by $(x_1, \ldots, x_n) \subset | |
R$ the subset of linear combinations $\sum r_i x_i$, where $r_i \in R$. This | |
is clearly an ideal, and in fact the smallest one containing all $x_i$. It is | |
called the ideal \textbf{generated} by $x_1, \ldots, x_n$. A | |
\textbf{principal ideal} $(x)$ is one generated by a single $x \in R$. | |
\begin{example} | |
Ideals generalize the notion of divisibility. Note that | |
in $\mathbb{Z}$, the set of elements divisible by $n \in \mathbb{Z}$ forms the | |
ideal $I = n\mathbb{Z} = (n)$. We shall see that every ideal in $\mathbb{Z}$ is | |
of this form: $\mathbb{Z}$ is a \emph{principal ideal domain.} | |
\end{example} | |
Indeed, one can think of an ideal as axiomatizing the notions that | |
``divisibility'' ought to satisfy. Clearly, if two elements are divisible by | |
something, then their sum and product should also be divisible by it. More | |
generally, if an element is divisible by something, then the product of that | |
element with anything else should also be divisible. In general, we will extend | |
(in the chapter on Dedekind domains) much of the ordinary arithmetic with | |
$\mathbb{Z}$ to arithmetic with \emph{ideals} (e.g. unique factorization). | |
\begin{example} | |
We saw in \cref{ringoffns} | |
that if $X$ is a set and $R$ a ring, then the set $R^X$ of functions $X \to R$ | |
is naturally a ring. If $Y \subset X$ is a subset, then the subset of functions | |
vanishing on $Y$ is an ideal. | |
\end{example} | |
\begin{exercise} | |
Show that the ideal $(2, 1 + \sqrt{-5}) \subset \mathbb{Z}[\sqrt{-5}]$ is not | |
principal. | |
\end{exercise} | |
\subsection{Operations on ideals} | |
There are a number of simple operations that one may do with ideals, which we | |
now describe. | |
\begin{definition} | |
The sum $I + J$ of two ideals $I, J \subset R$ is defined as the set of sums | |
\[ \left\{ x + y : x \in I, y \in J \right\}. \] | |
\end{definition} | |
\begin{definition} | |
The product $IJ$ of two ideals $I, J \subset R$ is defined as the smallest | |
ideal containing the products $xy$ for all $x \in I, y \in J$. This is just | |
the set | |
\[ \left\{ \sum x_i y_i : x_i \in I, y_i \in J \right\}. \] | |
\end{definition} | |
We leave the basic verification of properties as an exercise: | |
\begin{exercise} | |
Given ideals $I, J \subset R$, verify the following. | |
\begin{enumerate} | |
\item $I + J$ is the smallest ideal containing $I$ and $J$. | |
\item $IJ$ is contained in $I$ and $J$. | |
\item $I \cap J$ is an ideal. | |
\end{enumerate} | |
\end{exercise} | |
\begin{example} | |
In $\mathbb{Z}$, we have the following for any $m, n$. | |
\begin{enumerate} | |
\item $(m) + (n) = (\gcd\{ m, n \})$, | |
\item $(m)(n) = (mn)$, | |
\item $(m) \cap (n) = (\mathrm{lcm}\{ m, n \})$. | |
\end{enumerate} | |
\end{example} | |
\begin{proposition} | |
For ideals $I, J, K \subset R$, we have the following. | |
\begin{enumerate} | |
\item Distributivity: $I(J + K) = IJ + IK$. | |
\item $I \cap (J + K) = I \cap J + I \cap K$ if $I \supset J$ or $I \supset K$. | |
\item If $I + J = R$, $I \cap J = IJ$. | |
\end{enumerate} | |
\begin{proof} | |
1 and 2 are clear. For 3, note that $(I + J)(I \cap J) = I(I \cap J) | |
+ J(I \cap J) \subset IJ$. Since $IJ \subset I \cap J$, the result | |
follows. | |
\end{proof} | |
\end{proposition} | |
\begin{exercise} | |
There is a \emph{contravariant} functor $\mathbf{Rings} \to \mathbf{Sets}$ that | |
sends each ring to its set of ideals. Given a map $f: R \to S$ and an ideal $I | |
\subset S$, we define an ideal $f^{-1}(I) \subset R$; this defines the | |
functoriality. | |
This functor is not representable, as it does not send the initial object | |
in $\mathbf{Rings} $ to the | |
one-element set. We will later use a \emph{subfunctor} of this functor, the | |
$\spec$ construction, when we replace ideals with ``prime'' ideals. | |
\end{exercise} | |
\subsection{Quotient rings} | |
We next describe a procedure for producing new rings from old ones. | |
If $R$ is a ring and $I \subset R$ an ideal, then the quotient group $R/I$ | |
is a ring in its own right. If $a+I, b+I$ are two cosets, then the | |
multiplication is $(a+I)(b+I) = ab + I$. It is easy to check that this does | |
not depend on the coset representatives $a,b$. In other words, as mentioned | |
earlier, the arithmetic operations on $R$ \emph{factor} through the equivalence | |
relation defined by $I$. | |
As one easily checks, this becomes to a multiplication | |
\[ R/I \times R/I \to R/I \] | |
which is commutative and associative, and | |
whose identity element is $1+I$. | |
In particular, $R/I$ is a ring, under multiplication $(a+I)(b+I) = ab+I$. | |
\begin{definition} | |
$R/I$ is called the \textbf{quotient ring} by the ideal $I$. | |
\end{definition} | |
The process is analogous to quotienting a group by a normal subgroup: again, | |
the point is that the equivalence relation induced on the algebraic | |
structure---either the group or the ring---by the subgroup (or ideal)---is | |
compatible with the algebraic structure, which thus descends to the quotient. | |
The | |
reduction map $\phi \colon R \to R/I$ is a ring-homomorphism with a | |
\emph{universal | |
property}. | |
Namely, for any ring $B$, there is a map | |
\[ \hom(R/I, B) \to \hom(R, B) \] | |
on the hom-sets | |
by composing with the ring-homomorphism $\phi$; this map is injective and the | |
image consists of all homomorphisms $R \to B$ which vanish on $I$. | |
Stated alternatively, to map out of $R/I$ (into some ring $B$) is the same thing as mapping out of | |
$R$ while killing the ideal $I \subset R$. | |
This is best thought out for oneself, but here is the detailed justification. | |
The reason is that any map $R/I \to B$ pulls back to a map $R \to R/I \to B$ | |
which annihilates $I$ since $R \to R/I$ annihilates $I$. Conversely, if we have | |
a map | |
\[ f: R \to B \] | |
killing $I$, then we can define $R/I \to B$ by sending $a+I$ to $f(a)$; this is | |
uniquely defined since $f$ annihilates $I$. | |
\begin{exercise} | |
If $R$ is a commutative | |
ring, an element $e \in R$ is said to be \textbf{idempotent} if $e^2 = | |
e$. Define a covariant functor $\mathbf{Rings} \to \mathbf{Sets}$ sending a | |
ring to its idempotents. Prove that it is corepresentable. (Answer: the | |
corepresenting object is $\mathbb{Z}[X]/(X - X^2)$.) | |
\end{exercise} | |
\begin{exercise} | |
Show that the functor assigning to each ring the set of elements annihilated | |
by 2 is corepresentable. | |
\end{exercise} | |
\begin{exercise} | |
If $I \subset J \subset R$, then $J/I$ is an ideal of $R/I$, and there is a | |
canonical isomorphism | |
\[ (R/I)/(J/I) \simeq R/J. \] | |
\end{exercise} | |
\subsection{Zerodivisors} | |
Let $R$ be a commutative ring. | |
\begin{definition} | |
If $r \in R$, then $r$ is called a \textbf{zerodivisor} if there is $s \in R, s | |
\neq 0$ with $sr = 0$. Otherwise $r$ is called a \textbf{nonzerodivisor.} | |
\end{definition} | |
As an example, we prove a basic result on the zerodivisors in a polynomial ring. | |
\begin{proposition} | |
Let $A=R[x]$. Let $f=a_nx^n+\cdots +a_0\in A$. If there is a non-zero polynomial $g\in | |
A$ such that $fg=0$, then there exists $r\in R\smallsetminus\{0\}$ such that $f\cdot | |
r=0$. | |
\end{proposition} | |
So all the coefficients are zerodivisors. | |
\begin{proof} | |
Choose $g$ to be of minimal degree, with leading coefficient $bx^d$. We may assume | |
that $d>0$. Then $f\cdot b\neq 0$, lest we contradict minimality of $g$. We must have | |
$a_i g\neq 0$ for some $i$. To see this, assume that $a_i\cdot g=0$, then $a_ib=0$ for | |
all $i$ and then $fb=0$. Now pick $j$ to be the largest integer such that $a_jg\neq | |
0$. Then $0=fg=(a_0 + a_1x + \cdots a_jx^j)g$, and looking at the leading coefficient, | |
we get $a_jb=0$. So $\deg (a_jg)<d$. But then $f\cdot (a_jg)=0$, contradicting | |
minimality of $g$. | |
\end{proof} | |
\begin{exercise} | |
The product of two nonzerodivisors is a nonzerodivisor, and the product of two | |
zerodivisors is a zerodivisor. It is, however, not necessarily true that the | |
\emph{sum} of two zerodivisors is a zerodivisor. | |
\end{exercise} | |
\section{Further examples} | |
We now illustrate a few important examples of | |
commutative rings. The section is in large measure an advertisement for why | |
one might care about commutative algebra; nonetheless, the reader is | |
encouraged at least to skim this section. | |
\subsection{Rings of holomorphic functions} | |
The following subsection may be omitted without impairing understanding. | |
There is a fruitful analogy in number theory between the rings $\mathbb{Z}$ and | |
$\mathbb{C}[t]$, the latter being the polynomial ring over $\mathbb{C}$ in one | |
variable (\rref{polynomial}). Why are they analogous? Both of these rings have a theory of unique | |
factorization: that is, factorization into primes or irreducible polynomials. (In the | |
latter, the irreducible polynomials have degree one.) | |
Indeed we know: | |
\begin{enumerate} | |
\item Any nonzero integer factors as a product of primes (possibly times $-1$). | |
\item Any nonzero polynomial factors as a product of an element of | |
$\mathbb{C}^* =\mathbb{C} - \left\{0\right\}$ and polynomials of the form $t - | |
a, a \in \mathbb{C}$. | |
\end{enumerate} | |
There is another way of thinking of $\mathbb{C}[t]$ in terms of complex | |
analysis. This is equal to the ring of holomorphic functions on $\mathbb{C}$ | |
which are meromorphic at infinity. | |
Alternatively, consider the Riemann sphere $\mathbb{C} \cup \{ \infty\}$; then the ring $\mathbb{C}[t]$ | |
consists of meromorphic functions on the sphere whose poles (if any) are at | |
$\infty$. | |
This description admits generalizations. | |
Let $X$ be a | |
Riemann surface. (Example: take the complex numbers modulo a lattice, i.e. an | |
elliptic curve.) | |
Suppose that $x \in X$. Define $R_x$ to be the ring of meromorphic functions on $X$ | |
which are allowed poles only at $x$ (so are everywhere else holomorphic). | |
\begin{example} Fix the notations of the previous discussion. | |
Fix $y \neq x \in X$. Let $R_x$ be the ring of meromorphic functions on the | |
Riemann surface $X$ which are holomorphic on $X - \left\{x\right\}$, as before. | |
Then the collection of functions that vanish at $y$ forms an | |
\emph{ideal} in $R_x$. | |
There are lots of other ideals. For instance, fix two | |
points $y_0, y_1 \neq x$; we look at the ideal of $R_x$ that vanish at both $y_0, y_1$. | |
\end{example} | |
\textbf{For any Riemann surface $X$, the conclusion of Dedekind's theorem | |
(\rref{ded1}) applies. } In other | |
words, the ring $R_x$ as defined in the example admits unique factorization of | |
ideals. We shall call such rings \textbf{Dedekind domains} in the future. | |
\begin{example} Keep the preceding notation. | |
Let $f \in R_x$, nonzero. By definition, $f$ may have a pole at $x$, but no poles elsewhere. $f$ vanishes | |
at finitely many points $y_1, \dots, y_m$. When $X$ was the Riemann sphere, | |
knowing the zeros of $f$ told us something about $f$. Indeed, in this case | |
$f$ is just a | |
polynomial, and we have a nice factorization of $f$ into functions in $R_x$ that vanish only | |
at one point. In general Riemann surfaces, this | |
is not generally possible. This failure turns out to be very interesting. | |
Let $X = \mathbb{C}/\Lambda$ be an elliptic curve (for $\Lambda \subset | |
\mathbb{C}^2$ a lattice), and suppose $x = 0$. Suppose we | |
are given $y_1, y_2, \dots, y_m \in X$ that are nonzero; we ask whether there | |
exists a function $f \in R_x$ having simple zeros at $y_1, \dots, y_m$ and nowhere else. | |
The answer is interesting, and turns out to recover the group structure on the | |
lattice. | |
\begin{proposition} | |
A function $f \in R_x$ with simple zeros only at the $\left\{y_i\right\}$ exists if and only if $y_1 + y_2 + \dots + y_n = 0$ (modulo $\Lambda$). | |
\end{proposition} | |
So this problem of finding a function with specified zeros is equivalent to | |
checking that the specific zeros add up to zero with the group structure. | |
In any case, there might not be such a nice function, but we have at least an | |
ideal $I$ of functions that have zeros (not necessarily simple) at $y_1, \dots, | |
y_n$. This ideal has unique factorization into the ideals of functions | |
vanishing at $y_1$, functions vanishing at $y_2$, so on. | |
\end{example} | |
\subsection{Ideals and varieties} | |
We saw in the previous subsection that ideals can be thought of as the | |
vanishing of functions. This, like divisibility, is another interpretation, | |
which is particularly interesting in algebraic geometry. | |
Recall the ring $\mathbb{C}[t]$ of complex polynomials discussed in the | |
last subsection. More generally, if $R$ is a ring, we saw in | |
\rref{polynomial} that the set $R[t]$ of polynomials with coefficients | |
in $R$ | |
is a ring. This is a construction that | |
can be iterated to get a polynomial ring in several variables over $R$. | |
\begin{example} | |
Consider the polynomial ring $\mathbb{C}[x_1, \dots, x_n]$. Recall that before | |
we thought of the ring $\mathbb{C}[t]$ as a ring of meromorphic functions. | |
Similarly each element of the polynomial ring $\mathbb{C}[x_1, \dots, x_n]$ | |
gives a function $\mathbb{C}^n \to \mathbb{C}$; we can think of the polynomial | |
ring as sitting inside the ring of all functions $\mathbb{C}^n \to \mathbb{C}$. | |
A question you might ask: What are the ideals in this ring? One way to get an | |
ideal is to pick a point $x=(x_1, \dots, x_n) \in \mathbb{C}^n$; consider the | |
collection of all functions $f \in \mathbb{C}[x_1, \dots, x_n]$ which vanish on | |
$x$; by the usual argument, this is an ideal. | |
There are, of course, other ideals. More generally, if $Y \subset | |
\mathbb{C}^n$, consider the collection of polynomial functions $f: | |
\mathbb{C}^n \to \mathbb{C}$ such that $f \equiv 0$ on | |
$Y$. This is easily seen to be an ideal in the polynomial ring. We thus have a | |
way of taking a subset of $\mathbb{C}^n$ and producing an ideal. | |
Let $I_Y$ be the ideal corresponding to $Y$. | |
This construction is not injective. One can have $Y \neq Y'$ but $I_Y = I_{Y'}$. For instance, if $Y$ is dense in | |
$\mathbb{C}^n$, then $I_Y = (0)$, because the only way a continuous function on | |
$\mathbb{C}^n$ can vanish on $Y$ is for it to be zero. | |
There is a much closer connection in the other direction. You might ask whether | |
all ideals can arise in this way. The quick answer is no---not even when $n=1$. The ideal $(x^2) \subset \mathbb{C}[x]$ cannot be obtained | |
in this way. It is easy to see that the only way we could get this as $I_Y$ is | |
for $Y=\left\{0\right\}$, but $I_Y$ in this case is just $(x)$, not $(x^2)$. | |
What's going wrong in this example is that $(x^2)$ is not a \emph{radical} | |
ideal. | |
\end{example} | |
\begin{definition}\label{def-radical-ideal} | |
An ideal $I \subset R$ is \textbf{radical} if whenever $x^2 \in I$, then $x \in | |
I$. | |
\end{definition} | |
The ideals $I_Y$ in the polynomial ring are all radical. This is obvious. | |
You might now ask whether this is the only obstruction. We now state a theorem | |
that we will prove later. | |
\begin{theorem}[Hilbert's Nullstellensatz] If $I \subset \mathbb{C}[x_1, \dots, | |
x_n]$ is a radical ideal, then $I = I_Y$ for some $Y \subset \mathbb{C}^n$. In | |
fact, the canonical choice of $Y$ is the set of points where all the functions | |
in $Y$ vanish.\footnote{Such a subset is called an algebraic variety.} | |
\end{theorem} | |
This will be one of the highlights of the present course. But before we can | |
get to it, there is much to do. | |
\begin{exercise} | |
Assuming the Nullstellensatz, show that any \emph{maximal} ideal in the | |
polynomial ring $\mathbb{C}[x_1, \dots, x_n]$ is of the form | |
$(x_1-a_1, \dots, x_n-a_n)$ for $a_1, \dots, a_n \in \mathbb{C}$. An ideal of a | |
ring is called \textbf{maximal} if the only ideal that contains it is the | |
whole ring (and it itself is not the whole ring). | |
As a corollary, deduce that if $I \subset \mathbb{C}[x_1, \dots, x_n]$ is a | |
proper ideal (an ideal is called \textbf{proper} if it is not equal to the | |
entire ring), then there exists $(x_1, \dots, x_n) \in \mathbb{C}^n$ such that | |
every polynomial in $I$ vanishes on the point $(x_1, \dots, x_n)$. This is | |
called the \textbf{weak Nullstellensatz.} | |
\end{exercise} | |
\section{Modules over a commutative ring} | |
We will now establish some basic terminology about modules. | |
\subsection{Definitions} | |
Suppose $R$ is a commutative ring. | |
\begin{definition} | |
An \textbf{$R$-module $M$} is an abelian group $M$ with a map $R \times M \to | |
M$ (written $(a,m) \to am$) such that | |
\begin{enumerate}[\textbf{M} 1] | |
\item $(ab) m = a(bm)$ for $a,b \in R, m \in M$, i.e. there is an associative law. | |
\item $1m | |
= m$; the unit acts as the identity. | |
\item There are distributive laws | |
on both sides: | |
$(a+b)m = am + bm$ and $a(m+n) = am + an$ for $a,b \in R, \ m,n \in M$. | |
\end{enumerate} \end{definition} | |
Another definition can be given as follows. | |
\begin{definition} | |
If $M$ is an abelian group, $End(M)$ is the set of homomorphisms $f: M \to M$. | |
This can be made into a (noncommutative) \emph{ring}.\footnote{A | |
noncommutative ring is one satisfying all the usual axioms of a ring except | |
that multiplication is not required to be commutative.} Addition is defined pointwise, and | |
multiplication is by composition. The identity element is the identity | |
function $1_M$. | |
\end{definition} | |
We made the following definition earlier for commutative rings, but for | |
clarity we re-state it: | |
\begin{definition} | |
If $R, R'$ are rings (possibly noncommutative) then a function $f: R \to R'$ is a | |
\textbf{ring-homomorphism} or \textbf{morphism} if it is compatible with the | |
ring structures, i.e | |
\begin{enumerate} | |
\item $f(x+y) = f(x) + f(y)$ | |
\item $f(xy) = f(x)f(y)$ | |
\item $f(1) = 1$. | |
\end{enumerate} | |
\end{definition} | |
The last condition is not redundant because otherwise the zero map would | |
automatically be a homomorphism. | |
The alternative definition of a module is left to the reader in the following | |
exercise. | |
\begin{exercise} | |
If $R$ is a ring and $R \to End(M)$ a homomorphism, then $M$ is made into an | |
$R$-module, and vice versa. | |
\end{exercise} | |
\begin{example} | |
If $R$ is a ring, then $R$ is an $R$-module by multiplication on the left. | |
\end{example} | |
\begin{example} | |
A $\mathbb{Z}$-module is the same thing as an abelian group. | |
\end{example} | |
\begin{definition} | |
If $M$ is an $R$-module, a subset $M_0 \subset M$ is a \textbf{submodule} if it | |
is a subgroup (closed under addition and inversion) and is closed under | |
multiplication by elements of $R$, i.e. $aM_0 \subset M_0$ for $a \in R$. A | |
submodule is a module in its own right. If $M_0 \subset M$ is a submodule, | |
there is a commutative diagram: | |
\[ \xymatrix{ | |
R \times M_0 \ar[d] \ar[r] & M_0 \ar[d] \\ R \times M \ar[r] & M | |
}.\] | |
Here the horizontal maps are multiplication. | |
\end{definition} | |
\begin{example} | |
Let $R$ be a (\textbf{commutative}) ring; then an ideal in $R$ is the same thing as a | |
submodule of $R$. | |
\end{example} | |
\begin{example} | |
If $A$ is a ring, an $A$-algebra is an $A$-module in an obvious way. More | |
generally, if $A$ is a ring and $R$ is an $A$-algebra, any $R$-module becomes | |
an $A$-module by pulling back the multiplication map via $A \to R$. | |
\end{example} | |
Dual to submodules is the notion of a \emph{quotient module}, which we define | |
next: | |
\begin{definition} Suppose $M$ is an $R$-module and $M_0$ a | |
submodule. Then the abelian group $M/M_0$ (of cosets) is an $R$-module, | |
called the \textbf{quotient module} by $M_0$. | |
Multiplication is as follows. If | |
one has a coset $x + M_0 \in M/M_0$, one multiplies this by $a \in R$ to | |
get the coset $ax | |
+ M_0$. This does not depend on the coset representative. | |
\end{definition} | |
\subsection{The categorical structure on modules} | |
So far, we have talked about modules, but we have not discussed morphisms | |
between modules, and have yet to make the class of modules over a given ring | |
into a category. This we do next. | |
Let us thus introduce a few more basic notions. | |
\begin{definition} | |
Let $R$ be a ring. Suppose $M,N$ are $R$-modules. A map $f: M \to N$ | |
is a \textbf{module-homomorphism} if it preserves all the relevant structures. | |
Namely, it must be a homomorphism of abelian groups, $f(x+y) = f(x) + f(y)$, | |
and second it must | |
preserve multiplication: | |
$$f(ax) = af(x)$$ for $a \in R, x \in M$. | |
\end{definition} | |
A simple way of getting plenty of module-homomorphisms is simply to consider | |
multiplication by a fixed element of the ring. | |
\begin{example} | |
If $a \in R$, then multiplication by $a$ is a module-homomorphism $M | |
\stackrel{a}{\to} M$ for any $R$-module $M$.\footnote{When one considers | |
modules over noncommutative rings, this is no longer true.} Such homomorphisms | |
are called \textbf{homotheties.} | |
\end{example} | |
If $M \stackrel{f}{\to} N$ and $N \stackrel{g}{\to} P$ are | |
module-homomorphisms, their composite $M \stackrel{g \circ f}{\to} P$ clearly | |
is too. | |
Thus, for any commutative ring $R$, the class of $R$-modules and | |
module-homomorphisms forms a \textbf{category}. | |
\begin{exercise} | |
The initial object in this category is the zero module, and this is also the | |
final object. | |
In general, a category where the initial object and final object are the same | |
(that is, isomorphic) is called a \emph{pointed category.} The common object | |
is called the \emph{zero object.} In a pointed category $\mathcal{C}$, there is a morphism | |
$X \to Y$ for any two objects $X, Y \in \mathcal{C}$: if $\ast$ is the zero | |
object, then we can take $X \to \ast \to Y$. This is well-defined and is | |
called the \emph{zero morphism.} | |
One can easily show that the composition (on the left or the right) of a | |
zero morphism is a zero morphism (between a possibly different set of objects). | |
In the case of the category of modules, the zero object is clearly the zero | |
module, and the zero morphism $M \to N$ sends $m \mapsto 0$ for each $m \in M$. | |
\end{exercise} | |
\begin{definition} Let $f: M \to N$ be a module homomorphism. | |
In this case, the \textbf{kernel} $\ker f$ of $f$ is the set of elements $m | |
\in M$ with $f(m)=0$. This is | |
a submodule of $M$, as is easy to see. | |
The \textbf{image} $\im f$ of $f$ (the set-theoretic | |
image, i.e. the collection of all $f(x), x \in M$) is also a submodule of $N$. | |
The | |
\textbf{cokernel} of $f$ is defined by | |
\( N/\im(f). \) | |
\end{definition} | |
\begin{exercise} \label{univpropertykernel} | |
The universal property of the kernel is as follows. Let $M \stackrel{f}{\to } | |
N$ be a morphism with kernel $K \subset M$. Let $T \to M$ be a map. Then $T \to M$ factors through the | |
kernel $K \to M$ if and only if its composition with $f$ (a morphism $T \to N$) is zero. | |
That is, an arrow $T \to K$ exists in the diagram (where the dotted arrow | |
indicates we are looking for a map that need not exist) | |
\[ \xymatrix{ | |
& T \ar@{-->}[ld] \ar[d] \\ | |
K \ar[r] & M \ar[r]^f & N | |
}\] | |
if and only if the composite $T \to N$ is zero. | |
In particular, if we think of the hom-sets as abelian groups (i.e. | |
$\mathbb{Z}$-modules) | |
\[ \hom_R( T,K) = \ker\left( \hom_R(T, M) \to \hom_R(T, N) \right). \] | |
\end{exercise} | |
In other words, one may think of the kernel as follows. If $X | |
\stackrel{f}{\to} Y$ is a morphism, then the kernel $\ker(f)$ is the equalizer | |
of $f$ and the zero morphism $X \stackrel{0}{\to} Y$. | |
\begin{exercise} | |
What is the universal property of the cokernel? | |
\end{exercise} | |
\begin{exercise} \label{moduleunderlyingsetrepresentable} | |
On the category of modules, the functor assigning to each module $M$ its | |
underlying set is corepresentable (cf. \rref{corepresentable}). What | |
is the corepresenting object? | |
\end{exercise} | |
We shall now introduce the notions of \emph{direct sum} and \emph{direct | |
product}. Let $I$ be a set, and suppose that for each $i \in I$, we are given | |
an $R$-module $M_i$. | |
\begin{definition} | |
The \textbf{direct product} $\prod M_i$ is set-theoretically the cartesian product. It is given | |
the structure of an $R$-module by addition and multiplication pointwise on | |
each factor. | |
\end{definition} | |
\begin{definition} | |
The \textbf{direct sum} $\bigoplus_I M_i$ is the set of elements in the direct | |
product such that all but finitely many entries are zero. The direct sum is a | |
submodule of the direct product. | |
\end{definition} | |
\begin{example} \label{productcoproduct} | |
The direct product is a product in the category of modules, and the direct sum | |
is a coproduct. This is easy to verify: given maps $f_i: M \to M_i$, then we | |
get get a unique map $f: M \to \prod M_i$ by taking the product in the category | |
of sets. The case of a coproduct is dual: given maps $g_i: M_i \to N$, then we | |
get a map $\bigoplus M_i \to N$ by taking the \emph{sum} $g$ of the $g_i$: on a | |
family $(m_i) \in \bigoplus M_i$, we take $g(m_i) = \sum_I g_i(m_i)$; this is | |
well-defined as almost all the $m_i$ are zero. | |
\end{example} | |
\cref{productcoproduct} shows that the category of modules over a fixed | |
commutative ring has products and coproducts. In fact, the category of modules | |
is both complete and cocomplete (see \cref{completecat} for the definition). | |
To see this, it suffices to show that (by | |
\cref{coprodcoequalsufficeforcocomplete} and its dual) that this category | |
admits equalizers and coequalizers. | |
The equalizer of two maps | |
\[ M \stackrel{f,g}{\rightrightarrows} N \] | |
is easily checked to be the submodule of $M$ consisting of $m \in M$ such that | |
$f(m) = g(m)$, or, in other words, the kernel of $f-g$. The coequalizer of these two maps is the quotient module of $N$ | |
by the submodule $\left\{f(m) - g(m), m \in M\right\}$, or, in other words, | |
the cokernel of $f-g$. | |
Thus: | |
\begin{proposition} | |
If $R$ is a ring, the category of $R$-modules is complete and cocomplete. | |
\end{proposition} | |
\begin{example} | |
Note that limits in the category of $R$-modules are calculated in the same way | |
as they are for sets, but colimits are not. That is, the functor from | |
$R$-modules to $\mathbf{Sets}$, the forgetful functor, preserves limits but not | |
colimits. Indeed, we will see that the forgetful functor is a right adjoint | |
(\cref{freeadj}), which implies it preserves limits (by \cref{adjlimits}). | |
\end{example} | |
\subsection{Exactness} | |
Finally, we introduce the notion of \emph{exactness}. | |
\begin{definition} \label{exactness} | |
Let $f: M \to N$ be a morphism of $R$-modules. Suppose $g: N \to P$ is another morphism of | |
$R$-modules. | |
The pair of maps is a \textbf{complex} if $g \circ f = 0: M \to N \to P$. | |
This is equivalent to the condition that $\im(f) \subset \ker(g)$. | |
This complex is \textbf{exact} (or exact at $N$) if $\im(f) = \ker(g)$. | |
In other words, anything that is killed when mapped to $P$ actually comes from something in | |
$M$. | |
\end{definition} | |
We shall often write pairs of maps as sequences | |
\[ A \stackrel{f}{\to} B \stackrel{g}{\to} C \] | |
and say that the sequence is exact if the pair of maps is, as in | |
\rref{exactness}. A longer (possibly infinite) sequence of modules | |
\[ A_0 \to A_1 \to A_2 \to \dots \] | |
will be called a \textbf{complex} if each set of three | |
consecutive terms is a complex, and \textbf{exact} if it is exact at each step. | |
\begin{example} | |
The sequence $0 \to A \stackrel{f}{\to} B$ is exact if and only if the map $f$ | |
is injective. Similarly, $A \stackrel{f}{\to} B \to 0$ is exact if and only if | |
$f$ is surjective. Thus, $0 \to A \stackrel{f}{\to} B \to 0$ is exact if and | |
only if $f$ is an isomorphism. | |
\end{example} | |
One typically sees this definition applied to sequences of the form | |
\[ 0 \to M'\stackrel{f}{ \to} M \stackrel{g}{\to} M'' \to 0, \] | |
which, if exact, is called a \textbf{short exact sequence}. | |
Exactness here means that $f$ is injective, $g$ is surjective, and $f$ maps | |
onto the kernel of $g$. So $M''$ can be thought of as the quotient $M/M'$. | |
\begin{example} | |
Conversely, if $M$ is a module and $M' \subset M$ a submodule, then there is a | |
short exact sequence | |
\[ 0 \to M' \to M \to M/M' \to 0. \] | |
So every short exact sequence is of this form. | |
\end{example} | |
Suppose $F$ is a functor from the category of $R$-modules to the | |
category of $S$-modules, where $R, S$ are rings. Then: | |
\begin{definition} | |
\begin{enumerate} | |
\item $F$ is called \textbf{additive} if $F$ preserves direct sums. | |
\item $F$ is called \textbf{exact} if $F$ is additive and preserves exact sequences. | |
\item $F$ is called \textbf{left exact} if $F$ is additive and preserves exact sequences of the form | |
$0 \to M' \to M \to M''$. Equivalently, $F$ preserves kernels. | |
\item $F$ is \textbf{right exact} if $F$ is additive and $F$ preserves exact | |
sequences of the form $M' \to M \to M'' \to 0$, i.e. $F$ preserves cokernels. | |
\end{enumerate} | |
\end{definition} | |
The reader should note that much of homological algebra can be developed using the more | |
general setting of an \emph{abelian category,} which axiomatizes much of the | |
standard properties of the category of modules over a ring. Such a | |
generalization turns out to be necessary when many natural categories, such as | |
the category of chain complexes or the category of sheaves on a topological | |
space, are not naturally categories of modules. | |
We do not go into this here, cf. \cite{Ma98}. | |
A functor $F$ is exact if and only if it is both left and right exact. | |
This actually requires proof, though it is not hard. Namely, right-exactness implies that $F$ | |
preserves cokernels. Left-exactness implies that $F$ preserves kernels. $F$ | |
thus preserves images, as the image of a morphism is the kernel of its cokernel. | |
So if | |
\[ A \to B \to C \] | |
is a short exact sequence, then the kernel of the second map is equal to the | |
image of the first; we have just seen that this is preserved under $F$. | |
From this, one can check that left-exactness is equivalent to requiring that $F$ preserve | |
finite limits (as an additive functor, $F$ automatically preserves products, | |
and we have just seen that $F$ is left-exact iff it preserves kernels). | |
Similarly, right-exactness is equivalent to requiring that $F$ preserve | |
finite colimits. | |
So, in \emph{any} category with finite limits and colimits, we can talk about | |
right or left exactness of a functor, but the notion is used most often for | |
categories with an additive structure (e.g. categories of modules over a ring). | |
\begin{exercise} | |
Suppose whenever $0 \to A' \to A \to A'' \to 0$ is short exact, then $FA' \to | |
FA \to FA'' \to 0$ is exact. Prove that $F$ is right-exact. So we get a | |
slightly weaker criterion for right-exactness. | |
Do the same for left-exact functors. | |
\end{exercise} | |
\subsection{Split exact sequences} | |
Let $f: A \to B$ be a map of sets which is injective. Then there is a map $g: A | |
\to B$ such that the composite $g \circ f: A \stackrel{f}{\to} B | |
\stackrel{g}{\to} A$ is the identity. Namely, we define $g$ to be the inverse | |
of $f$ on $f(A)$ and arbitrarily on $B-f(A)$. | |
Conversely, if $f: A \to B$ admits an element $g: B \to A$ such that $g \circ f | |
= 1_A$, then $f$ is injective. This is easy to see, as any $a \in A$ can be | |
``recovered'' from $f(a)$ (by applying $g$). | |
In general, however, this observation does not generalize to arbitrary | |
categories. | |
\begin{definition} | |
Let $\mathcal{C}$ be a category. A morphism $A \stackrel{f}{\to} B$ is called a | |
\textbf{split injection} if there is $g: B \to A$ with $g \circ f = 1_A$. | |
\end{definition} | |
\begin{exercise}[General nonsense] | |
Suppose $f: A \to B$ is a split injection. Show that $f$ is a categorical monomorphism. | |
(Idea: the map $\hom(C,A) \to \hom(C,B)$ becomes a split injection of sets | |
thanks to $g$.) | |
\end{exercise} | |
\add{what is a categorical monomorphism? Maybe omit the exercise} | |
In the category of sets, we have seen above that \emph{any} monomorphism is a | |
split injection. This is not true in other categories, in general. | |
\begin{exercise} | |
Consider the morphism $\mathbb{Z} \to \mathbb{Z}$ given by multiplication by | |
2. Show that this is not a split injection: no left inverse $g$ can exist. | |
\end{exercise} | |
We are most interested in the case of modules over a ring. | |
\begin{proposition} | |
A morphism $f: A \to B$ in the category of $R$-modules is a split injection if | |
and only if: | |
\begin{enumerate} | |
\item $f$ is injective. | |
\item $f(A)$ is a direct summand in $B$. | |
\end{enumerate} | |
\end{proposition} | |
The second condition means that there is a submodule $B' \subset B$ such that | |
$B = B' \oplus f(A)$ (internal direct sum). In other words, $B = B' + f(A)$ | |
and $B' \cap f(A) = \left\{0\right\}$. | |
\begin{proof} | |
Suppose the two conditions hold, and we have a module $B'$ which is a | |
complement to $f(A)$. | |
Then we define a left inverse | |
\[ B \stackrel{g}{\to} A \] | |
by letting $g|_{f(A)} = f^{-1}$ (note that $f$ becomes an \emph{isomorphism} | |
$A \to f(A)$) and $g|_{B'}=0$. It is easy to see that this is indeed a left | |
inverse, though in general not a right inverse, as $g$ is likely to be | |
non-injective. | |
Conversely, suppose $f: A \to B$ admits a left inverse $g: B \to A$. The usual | |
argument (as for sets) shows that $f$ is injective. The essentially new | |
observation is that $f(A) $ is a direct summand in $B$. To define the | |
complement, we take $\ker(g) \subset B$. | |
It is easy to see (as $g \circ f = 1_A$) that $\ker(g) \cap f(A) = | |
\left\{0\right\}$. Moreover, $\ker(g) +f(A)$ fills $B$: given $b \in B$, it is | |
easy to check that | |
\[ b - f(g(b)) \in \ker(g). \] | |
Thus we find that the two conditions are satisfied. | |
\end{proof} | |
\add{further explanation, exactness of filtered colimits} | |
\subsection{The five lemma} | |
The five lemma will be a useful tool for us in proving that maps are | |
isomorphisms. Often this argument is used in inductive proofs. Namely, we will | |
see that often ``long exact sequences'' (extending infinitely in one or both | |
directions) arise from short exact sequences in a natural way. In such | |
events, the five lemma | |
will allow us to prove that certain morphisms are isomorphisms by induction on | |
the dimension. | |
\begin{theorem} | |
Suppose given a commutative diagram | |
\[ \xymatrix{ | |
A \ar[d] \ar[r] & B \ar[d] \ar[r] & C \ar[d] \ar[r] & D \ar[d] \ar[r] & E \ar[d] \\ | |
A' \ar[r] & B' \ar[r] & C' \ar[r] & D' \ar[r] & E' | |
}\] | |
such that the rows are exact and the four vertical maps $A \to A', B \to B', D | |
\to D', E \to E'$ are isomorphisms. Then $C \to C'$ is an isomorphism. | |
\end{theorem} | |
This is the type of proof that goes by the name of ``diagram-chasing,'' and | |
is best thought out visually for oneself, even though we give a complete proof. | |
\begin{proof} | |
We have the diagram | |
\[ | |
\xymatrix{ | |
A \ar[r]^k \ar[d]^\a & B \ar[r]^l \ar[d]^\b | |
& C \ar[r]^m \ar[d]^g & D \ar[r]^n \ar[d]^\d & E \ar[d]^\e \\ | |
F \ar[r]_p & G \ar[r]_q & H \ar[r]_r & I \ar[r]_s & J | |
} | |
\] | |
where the rows are exact at $B, C, D, G, H, I$ and the squares commute. In | |
addition, suppose that $\a, \b, \d, \e$ are isomorphisms. We will show that | |
$\g$ is an isomorphism. | |
\emph{We show that $\g$ is surjective:} | |
Suppose that $h \in H$. Since $\d$ is surjective, there exists an element | |
$d \in D$ such that $r(h) = \d(d) \in I$. | |
By the commutativity of the rightmost square, $s(r(h)) = \e(n(d))$. | |
The exactness at $I$ means that $\im r = \ker s$, so hence | |
$\e(n(d)) = s(r(h)) = 0$. Because $\e$ is injective, $n(d) = 0$. | |
Then $d \in \ker(n) = \im(m)$ by exactness at $D$. | |
Therefore, there is some $c \in C$ such that $m(c) = d$. | |
Now, $\d(m(c)) = \d(d) = r(h)$ and by the commutativity of squares, | |
$\d(m(c)) = r(\g(c))$, so therefore $r(\g(c)) = r(h)$. Since $r$ is a | |
homomorphism, $r(\g(c) - h) = 0$. Hence $\g(c) - h \in \ker r = \im q$ by | |
exactness at $H$. | |
Therefore, there exists $g \in G$ such that $q(g) = \g(c) - h$. | |
$\b$ is surjective, so there is some $b \in B$ such that $\b(b) = g$ and hence | |
$q(\b(b)) = \g(c) - h$. By the commutativity of squares, | |
$q(\b(b)) = \g(l(b)) = \g(c) - h$. Hence | |
$h = \g(c) - \g(l(b)) = \g(c - l(b))$, and therefore $\g$ is surjective. | |
So far, we've used that $\b$ and $\g$ are surjective, $\e$ is injective, and | |
exactness at $D$, $H$, $I$. | |
\emph{We show that $\g$ is injective:} | |
Suppose that $c \in C$ and $\g(c) = 0$. | |
Then $r(\g(c)) = 0$, and by the commutativity of squares, | |
$\d(m(c)) = 0$. Since $\d$ is injective, $m(c) = 0$, so | |
$c \in \ker m = \im l$ by exactness at $C$. | |
Therefore, there is $b \in B$ such that $l(b) = c$. | |
Then $\g(l(b)) = \g(c) = 0$, and by the commutativity of squares, | |
$q(\b(b)) = 0$. Therefore, $\b(b) \in \ker q$, and by exactness at $G$, | |
$\b(b) \in \ker q = \im p$. | |
There is now $f \in F$ such that $p(f) = \b(b)$. Since $\a$ is surjective, this | |
means that there is $a \in A$ such that $f = \a(a)$, so then | |
$\b(b) = p(\a(a))$. By commutativity of squares, | |
$\b(b) = p(\a(a)) = \b(k(a))$, and hence $\b(k(a) - b) = 0$. | |
Since $\b$ is injective, we have $k(a) -b = 0$, so $k(a) = b$. | |
Hence $b \in \im k = \ker l$ by commutativity of squares, so $l(b) = 0$. | |
However, we defined $b$ to satisfy $l(b) = c$, so therefore $c = 0$ and hence | |
$\g$ is injective. | |
Here, we used that $\a$ is surjective, $\b, \d$ are injective, and exactness at | |
$B, C, G$. | |
Putting the two statements together, we see that $\g$ is both surjective and | |
injective, so $\g$ is an isomorphism. We only used that $\b, \d$ are | |
isomorphisms and that $\a$ is surjective, $\e$ is injective, so we can slightly | |
weaken the hypotheses; injectivity of $\a$ and surjectivity of $\e$ were | |
unnecessary. | |
\end{proof} | |
\section{Ideals} | |
The notion of an \emph{ideal} has already been defined. Now we will introduce additional terminology related to the theory of ideals. | |
\subsection{Prime and maximal ideals} | |
Recall that the notion of an ideal generalizes that of divisibility. In | |
elementary number theory, though, one finds that questions of divisibility | |
basically reduce to questions about primes. | |
The notion of a ``prime ideal'' is intended to generalize the familiar idea of a prime | |
number. | |
\begin{definition} | |
An ideal $I \subset R$ is said to be \textbf{prime} if | |
\begin{enumerate}[\textbf{P} 1] | |
\item $1 \notin I$ (by convention, 1 is not a prime number) | |
\item If $xy \in I$, either $x \in I$ or $y \in I$. | |
\end{enumerate} | |
\end{definition} | |
\begin{example} | |
\label{integerprimes} | |
If $R = \mathbb{Z}$ and $p \in R$, then $(p) \subset \mathbb{Z}$ is a prime ideal iff $p$ or $-p$ is a | |
prime number in $\mathbb{N}$ or if $p$ is zero. | |
\end{example} | |
If $R$ is any commutative ring, there are two obvious ideals. These obvious | |
ones are the zero ideal $(0)$ | |
consisting only of the zero element, and the unit element $(1)$ consisting of all of | |
$R$. | |
\begin{definition} \label{maximalideal} | |
An ideal $I \subset R$ is called \textbf{maximal}\footnote{Maximal with | |
respect to not being the unit ideal.} if | |
\begin{enumerate}[\textbf{M} 1] | |
\item $1 \notin I$ | |
\item Any larger ideal contains $1$ (i.e., is all of $R$). | |
\end{enumerate} | |
\end{definition} | |
So a maximal ideal is a maximal element in the partially ordered set of proper | |
ideals (an ideal is \textbf{proper} if it does not contain 1). | |
\begin{exercise} | |
Find the maximal ideals in $\mathbb{C}[t]$. | |
\end{exercise} | |
\begin{proposition} | |
A maximal ideal is prime. | |
\end{proposition} | |
\begin{proof} | |
First, a maximal ideal does not contain 1. | |
Let $I \subset R$ be a maximal ideal. | |
We need to show that if $xy \in I$, | |
then one of $x,y \in I$. If $x \notin I$, then $(I,x) = I + (x)$ (the ideal | |
generated by $I$ and $x$) strictly contains $I$, so by maximality contains | |
$1$. In particular, $1 \in I+(x)$, so we can write | |
\[ 1 = a + xb \] | |
where $a \in I, b \in R$. Multiply both sides by $y$: | |
\[ y = ay + bxy. \] | |
Both terms on the right here are in $I$ ($a \in I$ and $xy \in I$), so we find | |
that $y \in I$. | |
\end{proof} | |
Given a ring $R$, what can we say about the collection of ideals in $R$? | |
There | |
are two obvious ideals in $R$, namely $(0)$ and $ (1)$. These are the same if and | |
only if $0=1$, i.e. $R$ is the zero ring. | |
So for any nonzero commutative ring, we have at least two distinct ideals. | |
Next, we show that maximal ideals always \emph{do} exist, except in the case | |
of the zero ring. | |
\begin{proposition} \label{anycontainedinmaximal} | |
Let $R$ be a commutative ring. Let $I \subset R$ be a proper ideal. Then $I$ | |
is contained in a maximal ideal. | |
\end{proposition} | |
\begin{proof} | |
This requires the axiom of choice in the form of Zorn's lemma. Let | |
$P$ be the collection of all ideals $J \subset R$ such that $I | |
\subset J$ and $J \neq R$. Then $P$ is a poset with respect to inclusion. $P$ is | |
nonempty because it contains $I$. Note that given a (nonempty) linearly ordered | |
collection of ideals $J_{\alpha} \in P$, the union $\bigcup J_{\alpha} \subset | |
R$ is an ideal: this is easily seen in view of the linear ordering (if $x,y | |
\in \bigcup J_{\alpha}$, then both $x,y$ belong to some $J_{\gamma}$, so $x+y | |
\in J_{\gamma}$; multiplicative closure is even easier). The union is not all | |
of $R$ because it does not contain $1$. | |
This implies that $P$ has a maximal element by Zorn's lemma. This maximal element may | |
be called $\mathfrak{M}$; it's a proper element containing $I$. I claim that | |
$\mathfrak{M}$ is a maximal ideal, because if it were contained in a larger | |
ideal, that would be in $P$ (which cannot happen by maximality) unless it were all of $R$. | |
\end{proof} | |
\begin{corollary} | |
Let $R $ be a nonzero commutative ring. Then $R$ has a maximal ideal. | |
\end{corollary} | |
\begin{proof} | |
Apply the lemma to the zero ideal. | |
\end{proof} | |
\begin{corollary} | |
Let $R$ be a nonzero commutative ring. Then $x \in R$ is invertible if and | |
only if it belongs to no maximal ideal $\mathfrak{m} \subset R$. | |
\end{corollary} | |
\begin{proof} | |
Indeed, $x$ is invertible if and only if $(x) = 1$. That is, if and only if | |
$(x)$ is not a proper ideal; now \rref{anycontainedinmaximal} | |
finishes the argument. | |
\end{proof} | |
\subsection{Fields and integral domains} | |
Recall: | |
\begin{definition} | |
A commutative ring $R$ is called a \textbf{field} if $1 \neq 0$ and for every $x \in R - | |
\left\{0\right\}$ there exists an \textbf{inverse} $x^{-1} \in R$ such that $xx^{-1} = | |
1$. | |
\end{definition} | |
This condition has an obvious interpretation in terms of ideals. | |
\begin{proposition} | |
A commutative ring with $1 \neq 0$ is a field iff it has only the two ideals $(1), | |
(0)$. | |
\end{proposition} | |
Alternatively, a ring is a field if and only if $(0)$ is a maximal ideal. | |
\begin{proof} | |
Assume $R$ is a field. Suppose $I \subset R$. If $I \neq (0)$, then there is | |
a nonzero $x \in I$. Then there is an inverse $x^{-1}$. We have $x^{-1} x =1 | |
\in I$, so $I = (1)$. | |
In a field, there is thus no room for ideals other than $(0)$ and $(1)$. | |
To prove the converse, assume every ideal of $R$ is $(0)$ or $(1)$. Then for | |
each $x \in R$, $(x) = (0)$ or $(1)$. If $x \neq 0$, the first cannot happen, so | |
that means that the ideal generated by $x$ is the unit ideal. So $1$ is a | |
multiple of $x$, implying that $x$ has a multiplicative inverse. | |
\end{proof} | |
So fields also have an uninteresting ideal structure. | |
\begin{corollary} \label{maximalfield} | |
If $R$ is a ring and $I \subset R$ is an ideal, then $I$ is maximal if and only | |
if $R/I$ is a field. | |
\end{corollary} | |
\begin{proof} | |
The basic point here is that there is a bijection between the ideals of $R/I$ | |
and ideals of $R$ containing $I$. | |
Denote by $\phi: R \to R/I$ the reduction map. There is a | |
construction mapping ideals of $R/I$ to ideals of $R$. This sends an ideal in | |
$R/I$ to | |
its inverse image. This is easily seen to map to ideals of $R$ containing $I$. | |
The map from ideals of $R/I$ to ideals of $R$ containing $I$ is a bijection, | |
as one checks easily. | |
It follows that $R/I$ is a field precisely if | |
$R/I$ has precisely two ideals, i.e. precisely if there are precisely two | |
ideals in $R$ containing $I$. These ideals must be $(1)$ and $I$, so this | |
holds if and only if $I$ is maximal. | |
\end{proof} | |
There is a similar characterization of prime ideals. | |
\begin{definition} | |
A commutative ring $R$ is an \textbf{integral domain} if for all $ x,y \in R$, | |
$x \neq 0 $ and $y \neq 0$ imply $xy \neq 0$. | |
\end{definition} | |
\begin{proposition} \label{primeifdomain} | |
An ideal $I \subset R$ is prime iff $R/I$ is a domain. | |
\end{proposition} | |
\begin{exercise} | |
Prove \rref{primeifdomain}. | |
\end{exercise} | |
Any field is an integral domain. This is because in a field, nonzero elements | |
are invertible, and the product of two invertible elements is invertible. This | |
statement translates in ring theory to the statement that a maximal ideal is | |
prime. | |
Finally, we include an example that describes what \emph{some} of the prime | |
ideals in a polynomial ring look like. | |
\begin{example} | |
Let $R$ be a ring and $P$ a prime ideal. We claim that $PR[x] \subset R[x]$ is a | |
prime ideal. | |
Consider the map $\tilde{\phi}:R[x]\rightarrow(R/P)[x]$ with | |
$\tilde{\phi}(a_0+\cdots+a_nx^n)=(a_0+P)+\cdots+(a_n+P)x^n$. This is clearly | |
a homomorphism because $\phi:R\rightarrow R/P$ is, and its kernel consists | |
of those polynomials $a_0+\cdots+a_nx^n$ with $a_0,\ldots,a_n\in P$, which is | |
precisely $P[x]$. Thus $R[x]/P[x]\simeq (R/P)[x]$, which is an integral domain | |
because $R/P$ is an integral domain. Thus $P[x]$ is a prime ideal. | |
However, if | |
$P$ is a maximal ideal, then $P[x]$ is never a maximal ideal because the ideal | |
$P[x]+(x)$ (the polynomials with constant term in $P$) always strictly contains | |
$P[x]$ (because if $x\in P[x]$ then $1\in P$, which is impossible). Note | |
that $P[x]+(x)$ is the kernel of the composition of $\tilde{\phi}$ with | |
evaluation at 0, i.e $(\text{ev}_0\circ\tilde{\phi}):R[x]\rightarrow R/P$, | |
and this map is a surjection and $R/P$ is a field, so that $P[x]+(x)$ is | |
the maximal ideal in $R[x]$ containing $P[x]$. | |
\end{example} | |
\begin{exercise} \label{quotfld1} | |
Let $R$ be a domain. Consider the set of formal quotients $a/b, a, b \in R$ | |
with $b \neq 0$. Define addition and multiplication using usual rules. Show | |
that the resulting object $K(R)$ is a ring, and in fact a \emph{field}. The | |
natural map $R \to K(R)$, $r \to r/1$, has a universal property. If $R | |
\hookrightarrow L$ is an injection of $R$ into a field $L$, then there is a | |
unique morphism $K(R) \to L$ of fields extending $R \to L$. This construction | |
will be generalized when we consider \emph{localization.} | |
This construction is called the \textbf{quotient field.} | |
Note that a non-injective map $R\to L$ will \emph{not} factor through the | |
quotient field! | |
\end{exercise} | |
\begin{exercise}\label{Jacobson} | |
Let $R$ be a commutative ring. Then the \textbf{Jacobson radical} of $R$ is | |
the intersection $\bigcap \mathfrak{m}$ of all maximal ideals $\mathfrak{m} | |
\subset R$. Prove that an element $x$ is in the Jacobson radical if and only | |
if $1 - yx$ is invertible for all $y \in R$. | |
\end{exercise} | |
\subsection{Prime avoidance} | |
The following fact will come in handy occasionally. We will, for instance, use | |
it much later to show that an ideal consisting of zerodivisors on a module $M$ is | |
contained in associated prime. | |
\begin{theorem}[Prime avoidance] \label{primeavoidance} | |
Let $I_1,\dots, I_n \subset R$ be ideals. Let $A\subset R$ be a subset which is closed | |
under addition and multiplication. Assume that at least $n-2$ of the ideals are | |
prime. If $A\subset I_1\cup \cdots \cup I_n$, then $A\subset I_j$ for some $j$. | |
\end{theorem} | |
The result is frequently used in the following specific case: if an ideal $I$ | |
is contained in a finite union $\bigcup \mathfrak{p}_i$ of primes, then $I | |
\subset \mathfrak{p}_i$ for some $i$. | |
\begin{proof} | |
Induct on $n$. If $n=1$, the result is trivial. The case $n=2$ is an easy argument: if | |
$a_1\in A\smallsetminus I_1$ and $a_2\in A\smallsetminus I_2$, then $a_1+a_2\in | |
A\smallsetminus (I_1\cup I_2)$. | |
Now assume $n\ge 3$. We may assume that for each $j$, $A\not\subset I_1\cup \cdots | |
\cup \hat I_j\cup \cdots I_n$.\footnote{The hat means omit $I_j$.} Fix an element | |
$a_j\in A\smallsetminus (I_1\cup \cdots \cup \hat I_j\cup \cdots I_n)$. Then this | |
$a_j$ must be contained in $I_j$ since $A\subset \bigcup I_j$. Since $n\ge 3$, one | |
of the $I_j$ must be prime. We may assume that $I_1$ is prime. Define | |
$x=a_1+a_2a_3\cdots a_n$, which is an element of $A$. Let's show that $x$ avoids | |
\emph{all} of the $I_j$. If $x\in I_1$, then $a_2a_3\cdots a_n\in I_1$, which | |
contradicts the fact that $a_i\not\in I_j$ for $i\neq j$ and that $I_1$ is prime. If | |
$x\in I_j$ for $j\ge 2$. Then $a_1\in I_j$, which contradicts $a_i\not\in I_j$ for | |
$i\neq j$. | |
\end{proof} | |
\subsection{The Chinese remainder theorem} | |
Let $m,n$ be relatively prime integers. Suppose $a, b \in \mathbb{Z}$; then | |
one can show that the two congruences $x \equiv a \mod m$ | |
and $x \equiv b \mod n$ can be solved simultaneously in $x \in \mathbb{Z}$. | |
The solution is unique, moreover, modulo $mn$. | |
The Chinese remainder theorem generalizes this fact: | |
\begin{theorem}[Chinese remainder theorem] Let $I_1, \dots I_n$ be ideals in a | |
ring $R$ which satisfy $I_i + I_j = R$ for $i \neq j$. Then we have $I_1 \cap | |
\dots \cap I_n = I_1 \dots I_n$ and the morphism of rings | |
\[ R \to \bigoplus R/I_i \] | |
is an epimorphism with kernel $I_1 \cap \dots \cap I_n$. | |
\end{theorem} | |
\begin{proof} | |
First, note that for any two ideals $I_1$ and $I_2$, we | |
have $I_1I_2\subset I_1\cap I_2$ and $(I_1+I_2)(I_1\cap I_2)\subset | |
I_1I_2$ (because any element of $I_1+I_2$ multiplied by any element of | |
$I_1\cap I_2$ will clearly be a sum of products of elements from both $I_1$ | |
and $I_2$). Thus, if $I_1$ and $I_2$ are coprime, i.e. $I_1+I_2=(1)=R$, | |
then $(1)(I_1\cap I_2)=(I_1\cap I_2)\subset I_1I_2\subset I_1\cap I_2$, | |
so that $I_1\cap I_2=I_1I_2$. This establishes the result for $n=2$. | |
If the | |
ideals $I_1,\ldots,I_n$ are pairwise coprime and the result holds for $n-1$, | |
then $$\bigcap_{i=1}^{n-1} I_i=\prod_{i=1}^{n-1}I_i.$$ Because $I_n+I_i=(1)$ | |
for each $1\leq i\leq n-1$, there must be $x_i\in I_n$ and $y_i\in I_i$ such | |
that $x_i+y_i=1$. Thus, $z_n=\prod_{i=1}^{n-1}y_i=\prod_{i=1}^{n-1}(1-x_i)\in | |
\prod_{i=1}^{n-1} I_i$, and clearly $z_n+I_n=1+I_n$ since each $x_i\in | |
I_n$. Thus $I_n+\prod_{i=1}^{n-1}I_i=I_n+\bigcap_{i=1}^{n-1}I_i=(1)$, | |
and we can now apply the $n=2$ case to conclude that $\bigcap_{i=1}^n | |
I_i=\prod_{i=1}^n I_i$. | |
Note that for any $i$, we can construct a $z_i$ | |
with $z_i\in I_j$ for $j\neq i$ and $z_i+I_i=1+I_i$ via the same procedure. | |
Define $\phi:R\rightarrow\bigoplus R/I_i$ | |
by $\phi(a)=(a+I_1,\ldots,a+I_n)$. The kernel of $\phi$ is | |
$\bigcap_{i=1}^n I_i$, because $a+I_i=0+I_i$ iff $a\in I_i$, so that | |
$\phi(a)=(0+I_1,\ldots,0+I_n)$ iff $a\in I_i$ for all $i$, that is, | |
$a\in\bigcap_{i=1}^n I_i$. Combined with our previous result, the kernel | |
of $\phi$ is $\prod_{i=1}^n I_i$. | |
Finally, recall that we constructed | |
$z_i\in R$ such that $z_i+I_i=1+I_i$, and $z+I_j=0+I_j$ for all $j\neq | |
i$, so that $\phi(z_i)=(0+I_1,\ldots,1+I_{i},\ldots,0+I_n)$. Thus, | |
$\phi(a_1z_1+\cdots+a_nz_n)=(a_1+I_1,\ldots,a_n+I_n)$ for all $a_i\in R$, | |
so that $\phi$ is onto. By the first isomorphism theorem, we have that | |
$R/I_1\cdots I_n\simeq \bigoplus_{i=1}^nR/I_i$. \\ | |
\end{proof} | |
\section{Some special classes of domains} | |
\subsection{Principal ideal domains} | |
\begin{definition} | |
A ring $R$ is a \textbf{principal ideal domain} or \textbf{PID} if $R \neq 0$, $R$ is not a | |
field, $R$ is a domain, and every ideal of $R$ is principal. | |
\end{definition} | |
These have the next simplest theory of ideals. | |
Each ideal is very simple---it's principal---though there might be a lot of ideals. | |
\begin{example} | |
$\mathbb{Z}$ is a PID. The only nontrivial fact to check here is that: | |
\begin{proposition} | |
Any nonzero ideal $I \subset \mathbb{Z}$ is principal. | |
\end{proposition} | |
\begin{proof} | |
If $I = (0)$, then this is obvious. Else there is $n \in I - | |
\left\{0\right\}$; we can assume $n>0$. Choose $n \in I$ as small as possible and | |
positive. Then I claim that the ideal $I$ is generated by $(n)$. Indeed, we have $(n) | |
\subset I$ obviously. If $m \in I$ is another integer, then divide $m$ by $n$, | |
to find $m = nb + r$ for $r \in [0, n)$. We find that $r \in I$ and $0 \leq r < | |
n$, so $r=0$, and $m$ is divisible by $n$. And $I \subset (n)$. | |
So $I = (n)$. | |
\end{proof} | |
\end{example} | |
A module $M$ is said to be \emph{finitely generated} if there exist elements | |
$x_1, \dots, x_n \in M$ such that any element of $M$ is a linear combination | |
(with coefficients in $R$) of the $x_i$. (We shall define this more formally | |
below.) | |
One reason that PIDs are so convenient is: | |
\begin{theorem}[Structure theorem] \label{structurePID} | |
If $M$ is a finitely generated module over a principal ideal domain $R$, then | |
$M$ is isomorphic to a direct sum | |
\[ M \simeq \bigoplus_{i=1}^n R/a_i, \] | |
for various $a_i \in R$ (possibly zero). | |
\end{theorem} | |
\add{at some point, the proof should be added. This is important!} | |
\subsection{Unique factorization domains} | |
The integers $\mathbb{Z}$ are especially nice because of the fundamental | |
theorem of arithmetic, which states that every integer has a unique | |
factorization into primes. This is not true for every integral domain. | |
\begin{definition} | |
An element of a domain $R$ is \textbf{irreducible} if it cannot be written | |
as the product of two non-unit elements of $R$. | |
\end{definition} | |
\begin{example} | |
Consider the integral domain $\mathbb{Z}[\sqrt{-5}]$. We saw earlier that | |
\[ | |
6 = 2 \cdot 3 = (1 + \sqrt{-5})(1 - \sqrt{-5}), | |
\] | |
which means that $6$ was written as the product of two non-unit elements in | |
different ways. $\mathbb{Z}[\sqrt{-5}]$ does not have unique factorization. | |
\end{example} | |
\begin{definition} \label{earlyUFD} | |
A domain $R$ is a \textbf{unique factorization domain} or \textbf{UFD} if every | |
non-unit $x \in R$ satisfies | |
\begin{enumerate} | |
\item $x$ can be written as a product $x = p_1 p_2 \cdots p_n$ of | |
irreducible elements $p_i \in R$ | |
\item if $x = q_1 q_2 \cdots q_m$ where $q_i \in R$ are irreducible | |
then the $p_i$ and $q_i$ are the same up to order and multiplication by units. | |
\end{enumerate} | |
\end{definition} | |
\begin{example} | |
$\mathbb{Z}$ is a UFD, while $\mathbb{Z}[\sqrt{-5}]$ is not. In fact, many of | |
our favorite domains have unique factorization. We will prove that all PIDs | |
are UFDs. In particular, in \rref{gaussianintegersareprincipal} and | |
\rref{polyringisprincipal}, we saw that $\mathbb{Z}[i]$ and $F[t]$ are PIDs, | |
so they also have unique factorization. | |
\end{example} | |
\begin{theorem} \label{PIDUFD} | |
Every principal ideal domain is a unique factorization domain. | |
\end{theorem} | |
\begin{proof} | |
Suppose that $R$ is a principal ideal domain and $x$ is an element of $R$. We | |
first demonstrate that $x$ can be factored into irreducibles. | |
If $x$ is a unit or an irreducible, then we are done. Therefore, we can assume | |
that $x$ is reducible, which means that $x = x_1 x_2$ for non-units | |
$x_1, x_2 \in R$. If there are irreducible, then we are again done, so we | |
assume that they are reducible and repeat this process. We need to show that | |
this process terminates. | |
Suppose that this process continued infinitely. Then we have an infinite | |
ascending chain of ideals, where all of the inclusions are proper: | |
$(x) \subset (x_1) \subset (x_{11}) \subset \cdots \subset R$. | |
We will show that this is impossible because any infinite ascending chain of | |
ideals $I_1 \subset I_2 \subset \cdots \subset R$ of a principal ideal domain | |
eventually becomes stationary, i.e. for some $n$, $I_k = I_n$ for $k \geq n$. | |
Indeed, let $I = \bigcup_{i=1}^\infty I_i$. This is an ideal, so it is | |
principally generated as $I = (a)$ for some $a$. Since $a \in I$, we must have | |
$a \in I_N$ for some $N$, which means that the chain stabilizes after $I_N$. | |
It remains to prove that this factorization of $x$ is unique. We induct on | |
the number of irreducible factors $n$ of $x$. If $n = 0$, then $x$ is a unit, | |
which has unique factorization up to units. Now, suppose that | |
$x = p_1 \cdots p_n = q_1 \cdots q_m$ for some $m \ge n$. Since $p_1$ divides | |
$x$, it must divide the product $q_1 \cdots q_m$ and by irreducibility, one of | |
the factors $q_i$. Reorder the $q_i$ so that $p_1$ divides $q_1$. However, | |
$q_1$ is irreducible, so this means that $p_1$ and $q_1$ are the same up to | |
multiplication by a unit $u$. Canceling $p_1$ from each of the two | |
factorizations, we see that $p_2 \cdots p_n = u q_2 \cdots q_m = q_2' \cdots | |
q_m$. By induction, this shows that the factorization of $x$ is unique up to | |
order and multiplication by units. | |
\end{proof} | |
\subsection{Euclidean domains} | |
A euclidean domain is a special type of principal ideal domain. In practice, | |
it will often happen that one has an explicit proof that a given domain is | |
euclidean, while it might not be so trivial to prove that it is a UFD without | |
the general implication below. | |
\begin{definition} | |
An integral domain $R$ is a \textbf{euclidean domain} if there is a function | |
$|\cdot |:R\to \mathbb \mathbb{Z}_{\geq 0}$ (called the norm) such that the following hold. | |
\begin{enumerate} | |
\item $|a|=0$ iff $a=0$. | |
\item For any nonzero $a,b\in R$ there exist $q,r\in R$ such that $b=aq+r$ and $|r|<|a|$. | |
\end{enumerate} | |
In other words, the norm is compatible with division with remainder. | |
\end{definition} | |
\begin{theorem}\label{EDPID} | |
A euclidean domain is a principal ideal domain. | |
\end{theorem} | |
\begin{proof} | |
Let $R$ be an euclidean domain, $I\subset R$ and ideal, and $b$ be the nonzero element of smallest norm in $I$. | |
Suppose $ a\in I$. Then we can write $ a = qb + r$ with $ 0\leq r < |b|$, but since $ b$ has minimal nonzero absolute value, $ r = 0$ and $ b|a$. Thus $ I=(b)$ is principal. | |
\end{proof} | |
As we will see, this implies that any euclidean domain admits \emph{unique | |
factorization.} | |
\begin{proposition} \label{polyringED} | |
Let $F$ be a field. Then the polynomial ring $F[t]$ is a euclidean domain. | |
In particular, it is a PID. | |
\end{proposition} | |
\begin{proof} | |
We define \add{} | |
\end{proof} | |
\begin{exercise} \label{gaussianintegersareprincipal} | |
Prove that $\mathbb{Z}[i]$ is principal. | |
(Define the norm as $N(a+ib) = a^2 + b^2$.) | |
\end{exercise} | |
\begin{exercise} \label{polyringisprincipal} | |
Prove that the polynomial ring $F[t]$ for $F$ a field is principal. | |
\end{exercise} | |
It is \emph{not} true that a PID is necessarily euclidean. Nevertheless, it | |
was shown in \cite{Gre97} that the converse is ``almost'' true. Namely, | |
\cite{Gre97} defines the notion of an \textbf{almost euclidean domain.} | |
A domain $R$ is almost euclidean if there is a function $d: R \to | |
\mathbb{Z}_{\geq 0}$ such that | |
\begin{enumerate} | |
\item $d(a) = 0$ iff $a = 0$. | |
\item $d(ab) \geq d(a)$ if $b \neq 0$. | |
\item If $a,b \in R - \left\{0\right\}$, then either $b \mid a$ or there is | |
$r \in (a,b)$ with $d(r)<d(b)$. | |
\end{enumerate} | |
It is easy to see by the same argument that an almost euclidean domain is a PID. | |
(Indeed, let $R$ be an almost euclidean domain, and $I \subset R$ a nonzero | |
ideal. Then choose $x \in I - \left\{0\right\}$ such that $d(x)$ is minimal among elements in | |
$I$. Then if $y \in I - \left\{0\right\}$, either $x \mid y$ or $(x,y) \subset I$ contains an | |
element with smaller $d$. The latter cannot happen, so the former does.) | |
However, in fact: | |
\begin{proposition}[\cite{Gre97}] \label{almosteuclidean} | |
A domain is a PID if and only if it is almost euclidean. | |
\end{proposition} | |
\begin{proof} | |
Indeed, let $R$ be a PID. Then $R$ is a UFD (\rref{PIDUFD}), so for any $x \in R$, | |
there is a factorization into prime elements, unique up to units. If $x$ | |
factors into $n$ elements, we define $d(x)=n$; we set $d(0)=0$. | |
The first two conditions for an almost euclidean domain are then evident. | |
Let $x = p_1 \dots p_m$ and $y = q_1 \dots q_n$ be two elements of $R$, | |
factored into irreducibles. Suppose $x \nmid y$. Choose a generator $b$ of the (principal) ideal $(x,y)$; then obviously $y | |
\mid b$ so $d(y) \leq d(b)$. But if $d(y) = d(b)$, then the | |
number of factors of $y$ and $b$ is the same, so $y \mid b$ would imply | |
that $y$ and $b$ are associates. This is a contradiction, and implies that | |
$d(y)<d(b)$. | |
\end{proof} | |
\begin{remark} | |
We have thus seen that a euclidean domain is a PID, and a PID is a UFD. Both | |
converses, however, fail. By Gauss's lemma (\rref{}), the | |
polynomial ring $\mathbb{Z}[X]$ has unique factorization, though the ideal | |
$(2, X)$ is not principal. | |
In \cite{Ca88}, it is shown that the ring $\mathbb{Z}[\frac{1+ | |
\sqrt{-19}}{2}]$ is a PID but not euclidean (i.e. there is \emph{no} euclidean | |
norm on it). | |
\end{remark} | |
According to \cite{Cl11}, sec. 8.3, \cref{almosteuclidean} actually goes back to Hasse | |
(and these norms are sometimes called ``Dedekind-Hasse norms''). | |
\section{Basic properties of modules} | |
\subsection{Free modules} | |
We now describe a simple way of constructing modules over a ring, and an | |
important class of modules. | |
\begin{definition} | |
\label{freemoduledef} | |
A module $M$ is \textbf{free} if it is isomorphic to $\bigoplus_I R$ for some | |
index set $I$. The cardinality of $I$ is called the \textbf{rank}. | |
\end{definition} | |
\begin{example} | |
$R$ is the simplest example of a free module. | |
\end{example} | |
Free modules have a \emph{universal property}. | |
Namely, recall that if $M$ is an $R$-module, then to give a homomorphism | |
\[ R \to M \] | |
is equivalent to giving an element $m \in M$ (the image of $1$). | |
By the universal product of the direct sum (which is the coproduct in the | |
category of modules), it follows that to give a map | |
\[ \bigoplus_I \to M \] | |
is the same as giving a map of \emph{sets} $I \to M$. | |
In particular: | |
\begin{proposition} \label{freeadj} | |
The functor $I \mapsto \bigoplus_I R$ from $\mathbf{Sets}$ to | |
$R$-modules is the \emph{left adjoint} to the forgetful functor from | |
$R$-modules to $\mathbf{Sets}$. | |
\end{proposition} | |
The claim now is that the notion of ``rank'' is well-defined for a free | |
module. To see this, we will have to use the notion | |
of a \emph{maximal ideal} (\rref{maximalideal}) and | |
\rref{maximalfield}. | |
Indeed, suppose | |
$\bigoplus_I R$ and $\bigoplus_J R$ are isomorphic; we must show that $I$ and | |
$J$ have the same cardinality. Choose a maximal ideal $\mathfrak{m} | |
\subset R$. Then, by applying the functor $M \to | |
M/\mathfrak{m}M$, we find that the $R/\mathfrak{m}$-\emph{vector spaces} | |
\[ \bigoplus_I R/\mathfrak{m}, \quad \bigoplus_J R/\mathfrak{m} \] | |
are isomorphic. By linear algebra, $I$ and $J$ have the same cardinality. | |
Free modules have a bunch of nice properties. The first is that it is very | |
easy to map out of a free module. | |
\begin{example} | |
Let $I$ be an indexing set, and $M$ an $R$-module. Then to give a morphism | |
\[ \bigoplus_I R \to M \] | |
is equivalent to picking an element of $M$ for each $i \in I$. Indeed, given | |
such a collection of elements $\left\{m_i\right\}$, we send the generator of $\bigoplus_I R$ with a 1 | |
in the $i$th spot and zero elsewhere to $m_i$. | |
\end{example} | |
\begin{example} | |
In a domain, every principal ideal (other than zero) is a free module of rank | |
one. | |
\end{example} | |
Another way of saying this is that the free module $\bigoplus_I R$ represents | |
the functor on modules sending $M$ to the \emph{set} $ M^I$. We have already seen a special case of this for $I$ a | |
one-element set (\rref{moduleunderlyingsetrepresentable}). | |
The next claim is that free modules form a reasonably large class of the | |
category of $R$-modules. | |
\begin{proposition} \label{freesurjection} | |
Given an $R$-module $M$, there is a free module $F$ and a surjection | |
\[ F \twoheadrightarrow M. \] | |
\end{proposition} | |
\begin{proof} | |
We let $F$ to be the free $R$-module on the elements $e_m$, one for each $m | |
\in M$. We define the map | |
\[ F \to M \] | |
by describing the image of each of the generators $e_m$: we just send each | |
$e_m$ to $m \in M$. It is clear that this map is surjective. | |
\end{proof} | |
We close by making a few remarks on matrices. | |
Let $M$ be a free module of rank $n$, and fix an isomorphism $M \simeq R^n$. | |
Then we can do linear algebra with $M$, even though we are working over a | |
ring and not necessarily a field, at least to some extent. | |
For instance, we can talk about $n$-by-$n$ matrices over the ring $R$, and | |
then each of them induces a transformation, i.e. a module-homomorphism, $M \to | |
M$; it is easy to see that every module-homomorphism between free modules is | |
of this form. Moreover, multiplication of matrices corresponds to composition | |
of homomorphisms, as usual. | |
\begin{example} Let us consider the question of when the transformation | |
induced by an $n$-by-$n$ matrix is invertible. The answer is similar to the | |
familiar one from linear algebra in the case of a field. Namely, the condition | |
is that the determinant be invertible. | |
Suppose that an $n \times n$ matrix $A$ over a ring $R$ is invertible. This | |
means that there exists $A^{-1}$ so that $A A^{-1} = I$, so hence | |
$1 = \det I = \det(A A^{-1}) = (\det A) (\det A^{-1})$, and therefore, | |
$\det A$ must be a unit in $R$. | |
Suppose instead that an $n \times n$ matrix $A$ over a ring $R$ has an | |
invertible determinant. Then, using Cramer's rule, we can actually construct | |
the inverse of $A$. | |
\end{example} | |
We next show that if $R$ is a commutative ring, the category of modules over | |
$R$ contains enough information to reconstruct $R$. This is a small part of the | |
story of \emph{Morita equivalence,} which we shall not enter into here. | |
\begin{example} | |
Suppose $R$ is a commutative ring, and let $\mathcal{C}$ be the category of | |
$R$-modules. The claim is that $\mathcal{C}$, as an \emph{abstract} category, | |
determines $R$. Indeed, the claim is that $R$ is canonically the ring of | |
endomorphisms of the identity functor $1_{\mathcal{C}}$. | |
Such an \emph{endomorphism} is given by a natural transformation | |
$\phi: 1_{\mathcal{C}} \to 1_{\mathcal{C}}$. In other words, one requires for | |
each $R$-module $M$, a homomorphism of $R$-modules $\phi_M : M \to M$ such that | |
if $f: M \to N$ is any homomorphism of modules, then there is a commutative | |
square | |
\[ \xymatrix{ | |
M \ar[d]^f \ar[r]^{\phi_M} & M \ar[d] \\ | |
N \ar[r]^{\phi_N} & N. | |
}\] | |
Here is a simple way of obtaining such endomorphisms. Given $r \in R$, we | |
consider the map $r: M \to m$ which just multiplies each element by $r$. This | |
is a homomorphism, and it is clear that it is natural in the above sense. There | |
is thus a map $R \to \mathrm{End}(1_\mathcal{C})$ (note that multiplication | |
corresponds to composition of natural transformations). | |
This map is clearly injective; different $r, s \in R$ lead to different natural | |
transformations (e.g. on the $R$-module $R$). | |
The claim is that \emph{any} natural transformation of $1_{\mathcal{C}}$ is | |
obtained in this way. | |
Namely, let $\phi: 1_{\mathcal{C}} \to 1_{\mathcal{C}}$ be such a natural | |
transformation. On the $R$-module $R$, $\phi$ must be multiplication by some | |
element $r \in R$ | |
(because $\hom_R(R, R)$ is given by such homotheties). | |
Consequently, one sees by drawing commutative diagrams that $\phi: R^{\oplus S} | |
\to R^{\oplus S}$ is of this form for any set $S$. So $\phi$ is multiplication | |
by $r$ on any free $R$-module. | |
Since any module $M$ is a quotient of a free module $F$, we can draw a diagram | |
\[ \xymatrix{ | |
F\ar[d] \ar[r]^{\phi_F} & F \ar[d] \\ | |
M \ar[r]^{\phi_M} & M. | |
}\] | |
Since the vertical arrows are surjective, we find that $\phi_F$ must be given | |
by multiplication by $r$ too. | |
\end{example} | |
\subsection{Finitely generated modules} | |
The notion of a ``finitely generated'' module is analogous to that of a | |
finite-dimensional vector space. | |
\begin{definition} | |
An $R$-module $M$ is \textbf{finitely generated} if there exists a surjection | |
$R^n \to M$ for some $n$. In other words, it has a finite number of elements | |
whose ``span'' contains $M$. | |
\end{definition} | |
The basic properties of finitely generated modules follow from the fact that | |
they are stable under extensions and quotients. | |
\begin{proposition} \label{exact-fingen} | |
Let $0 \to M' \to M \to M'' \to 0$ be an exact sequence. If $M', M''$ are | |
finitely generated, so is $M$. | |
\end{proposition} | |
\begin{proof} | |
Suppose $0\rightarrow | |
M'\stackrel{f}{\rightarrow}M\stackrel{g}{\rightarrow}M''\rightarrow0$ | |
is exact. Then $g$ is surjective, $f$ is injective, and | |
$\text{ker}(g)=\text{im}(f)$. Now suppose $M'$ is finitely generated, | |
say by $\{a_1,\ldots,a_s\}$, and $M''$ is finitely generated, say by | |
$\{b_1,\ldots,b_t\}$. Because $g$ is surjective, each $g^{-1}(b_i)$ is | |
non-empty. Thus, we can fix some $c_i\in g^{-1}(b_i)$ for each $i$. | |
For any | |
$m\in M$, we have $g(m)=r_1b_1+\cdots+r_tb_t$ for some $r_i\in R$ because | |
$g(m)\in M''$ and $M''$ is generated by the $b_i$. Thus $g(m)=r_1g(c_i)+\cdots | |
r_tg(c_t)=g(r_1c_1+\cdots+r_tc_t)$, and because $g$ is a homomorphism | |
we have $m-(r_1c_1+\cdots+r_tc_t)\in\text{ker}(g)=\text{im}(f)$. But | |
$M'$ is generated by the $a_i$, so the submodule $\text{im}(f)\subset | |
M$ is finitely generated by the $d_i=f(a_i)$. | |
Thus, any $m\in | |
M$ has $m-(r_1c_1+\cdots+r_tc_t)=r_{t+1}d_1+\cdots+r_{t+s}d_s$ | |
for some $r_1,\ldots,r_{t+s}$, thus $M$ is finitely generated by | |
$c_1,\ldots,c_t,d_1,\ldots,d_s$. \\ | |
\end{proof} | |
The converse is false. It is possible for finitely generated modules to have | |
submodules which are \emph{not} finitely generated. As we shall see in | |
\rref{noetherian}, this does not happen over \emph{noetherian} rings. | |
\begin{example} | |
Consider the ring $R=\mathbb{C}[X_1, X_2, \dots,]$ and the ideal $(X_1, X_2, | |
\dots)$. This ideal is a submodule of the finitely generated $R$-module $R$, | |
but it is not finitely generated. | |
\end{example} | |
\begin{exercise} | |
Show that a quotient of a finitely generated module is finitely generated. | |
\end{exercise} | |
\begin{exercise} | |
Consider a \emph{split} exact sequence $0 \to M' \to M \to M'' \to 0$. In this | |
case, show that if $M$ is finitely generated, so is $M'$. | |
\end{exercise} | |
\subsection{Finitely presented modules} | |
Over messy rings, the notion of a finitely presented module is often a good | |
substitute for that of a finitely generated one. In fact, we are going to see | |
(\rref{}), that there is a general method of reducing questions about finitely | |
presented modules over arbitrary rings to finitely generated modules over | |
finitely generated $\mathbb{Z}$-algebras. | |
Throughout, fix a ring $R$. | |
\begin{definition} | |
An $R$-module $M$ is \textbf{finitely presented} if there is an exact sequence | |
\[ R^m \to R^n \to M \to 0. \] | |
\end{definition} | |
The point of this definition is that $M$ is the quotient of a free module | |
$R^n$ by the ``relations'' given by the images of the vectors in $R^m$. | |
Since $R^m$ is finitely generated, $M$ can be represented via finitely many | |
generators \emph{and} finitely many relations. | |
The reader should compare this with the definition of a \textbf{finitely | |
generated} module; there we only require an exact sequence | |
\[ R^n \to M \to 0. \] | |
As usual, we establish the usual properties of finitely presented modules. | |
We start by showing that if a finitely presented module $M$ is generated by | |
finitely many elements, the ``module of relations'' among these generators is | |
finitely generated itself. The condition of finite presentation only states that | |
there is \emph{one} such set of generators such that the module of generators | |
is finitely generated. | |
\begin{proposition} | |
Suppose $M$ is finitely presented. Then if $R^m \twoheadrightarrow M$ is a | |
surjection, the kernel is finitely generated. | |
\end{proposition} | |
\begin{proof} Let $K$ be the kernel of $R^m \twoheadrightarrow M$. | |
Consider an exact sequence | |
\[ F' \to F \to M \to 0 \] | |
where $F', F$ are finitely generated and free, which we can do as $M$ is | |
finitely presented. | |
Draw a commutative and exact diagram | |
\[ | |
\xymatrix{ | |
& F' \ar[r] & F \ar[r] \ar@{-->}[d] & M \ar[r] \ar[d] & 0 \\ | |
0 \ar[r] & K \ar[r] & R^m \ar[r] & M \ar[r] & 0 | |
} | |
\] | |
The dotted arrow $F \to R^m$ exists as $F$ is projective. There is induced a | |
map $F' \to K$. | |
We get a commutative and exact diagram | |
\[ | |
\xymatrix{ | |
& F' \ar[r]\ar[d]^f & F \ar[r] \ar[d]^g & M \ar[r] \ar[d] & 0 \\ | |
0 \ar[r] & K \ar[r] & R^m \ar[r] & M \ar[r] & 0 | |
}, | |
\] | |
to which we can apply the snake lemma. There is an exact sequence | |
\[ 0 \to \coker(f) \to \coker(g) \to 0, \] | |
which gives an isomorphism $\coker(f) \simeq \coker(g)$. | |
However, $\coker(g)$ is finitely generated, as a quotient of $R^m$. | |
Thus $\coker(f)$ is too. | |
Since we have an exact sequence | |
\[ 0 \to \im(f) \to K \to \coker(f) \to 0, \] | |
and $\im(f)$ is finitely generated (as the image of a finitely generated | |
object, $F'$), we find by \rref{exact-fingen} that $\coker(f)$ is finitely generated. | |
\end{proof} | |
\begin{proposition} \label{exact-finpres} | |
Given an exact sequence | |
\[ 0 \to M' \to M \to M'' \to 0, \] | |
if $M', M''$ are finitely presented, so is $M$. | |
\end{proposition} | |
In general, it is not true that if $M$ is finitely presented, then $M'$ and | |
$M''$ are. For instance, it is possible that a submodule of the free, finitely | |
generated module $R$ (i.e. an ideal), might fail to be finitely generated. We | |
shall see in \rref{noetherian} that this does not happen over a | |
\emph{noetherian} ring. | |
\begin{proof} | |
Indeed, suppose we have exact sequences | |
\[ F_1' \to F_0' \to M' \to 0 \] | |
and | |
\[ F_1'' \to F_0'' \to M'' \to 0 \] | |
where the $F$'s are finitely generated and free. | |
We need to get a similar sequence for $M$. | |
Let us stack these into a diagram | |
\[ \xymatrix{ | |
& F_1' \ar[d] & & F_1'' \ar[d] \\ | |
& F_0' \ar[d] & & F_0'' \ar[d] \\ | |
0 \ar[r] & M' \ar[r] & M \ar[r] & M'' \ar[r] & 0 | |
}\] | |
However, now, using general facts about projective modules (\rref{}), we can | |
splice these presentations into a resolution | |
\[ F_1' \oplus F_1'' \to F_0' \oplus F_0'' \to M \to 0, \] | |
which proves the assertion. | |
\end{proof} | |
\begin{corollary} | |
The (finite) direct sum of finitely presented modules is finitely presented. | |
\end{corollary} | |
\begin{proof} | |
Immediate from \rref{exact-finpres} | |
\end{proof} | |
\subsection{Modules of finite length} | |
A much stronger condition on modules that of finite generation is that of \emph{finite | |
length}. Here, basically any operation one does will eventually terminate. | |
Let $R$ be a commutative ring, $M$ an $R$-module. | |
\begin{definition} | |
$M$ is \textbf{simple} if $M \neq 0$ and $M$ has no nontrivial submodules. | |
\end{definition} | |
\begin{exercise} | |
A torsion-free abelian group is never a simple $\mathbb{Z}$-module. | |
\end{exercise} | |
\begin{proposition} | |
$M$ is simple if and only if it is isomorphic to $R/\mathfrak{m}$ for $\mathfrak{m} \subset | |
R$ a maximal ideal. | |
\end{proposition} | |
\begin{proof} Let $M$ be simple. Then | |
$M$ must contain a cyclic submodule $Rx$ generated by some $x \in | |
M - \left\{0\right\}$. So it must contain a submodule isomorphic to $R/I$ | |
for some ideal $I$, and | |
simplicity implies that $M \simeq R/I$ for some $I$. If $I$ is not maximal, | |
say properly contained in $J$, | |
then we will get a nontrivial submodule $J/I$ of $R/I \simeq M$. Conversely, | |
it is easy to see | |
that $R/\mathfrak{m}$ is simple for $\mathfrak{m}$ maximal. | |
\end{proof} | |
\begin{exercise}[Schur's lemma] Let $f: M \to N$ be a module-homomorphism, | |
where $M, N$ are both simple. Then either $f =0$ or $f$ is an isomorphism. | |
\end{exercise} | |
\begin{definition} | |
$M$ is of \textbf{finite length} if there is a finite filtration $0 \subset M^0 | |
\subset \dots \subset M^n = M$ where each $M^i/M^{i-1}$ is simple. | |
\end{definition} | |
\begin{exercise} | |
Modules of finite length are closed under extensions (that is, if $0 \to M' | |
\to M \to M'' \to 0$ is an exact sequence, then if $M', M''$ are of finite | |
length, so is $M$). | |
\end{exercise} | |
In the next result (which will not be used in this chapter), we shall use the | |
notions of a \emph{noetherian} and an \emph{artinian} module. These notions | |
will be developed at length in \cref{chnoetherian}, and we refer the reader | |
there for more explanation. | |
A module is \emph{noetherian} if every ascending chain $M_1 \subset M_2 \subset | |
\dots$ of submodules stabilizes, and it is \emph{artinian} if every descending chain | |
stabilizes. | |
\begin{proposition} | |
$M$ is finite length iff $M$ is both noetherian and artinian. | |
\end{proposition} | |
\begin{proof} | |
Any simple module is obviously both noetherian and artinian: there are two | |
submodules. So if $M$ is finite length, then the finite filtration with simple | |
quotients implies that $M$ is noetherian and artinian, since these two | |
properties are stable under extensions (\rref{exactnoetherian} | |
and \rref{exactartinian} of \rref{noetherian}). | |
Suppose $M \neq 0$ is noetherian and artinian. Let $M_1 \subset M$ be a minimal | |
nonzero submodule, which exists as $M$ is artinian. This is necessarily simple. Then we have a filtration | |
\[ 0 = M_0 \subset M_1. \] | |
If $M_1 = M$, then the filtration goes up to $M$, and we have that $M$ is of | |
finite length. If not, find a submodule $M_2$ that contains $M_1$ and is | |
minimal among submodules containing $M_1$; then the quotient | |
$M_2/M_1$ is simple. We have the filtration | |
\[ 0 = M_0 \subset M_1 \subset M_2, \] | |
which we can keep continuing until at some point we reach $M$. Note that since | |
$M$ is noetherian, we cannot continue this strictly ascending chain forever. | |
\end{proof} | |
\begin{exercise} | |
In particular, any submodule or quotient module of a finite length module is | |
of finite length. Note that the analog is not true for finitely generated | |
modules unless the ring in question is noetherian. | |
\end{exercise} | |
Our next goal is to show that the length of a filtration of a module with | |
simple quotients is well-defined. | |
For this, we need: | |
\begin{lemma} \label{simplefiltrationint} | |
Let $0 = M_0 \subset M_1 \subset \dots \subset M_n = M$ be a filtration of | |
$M$ with simple quotients. Let $N \subset M$. Then the filtration | |
$0 = M_0 \cap N \subset M_1 \cap N \subset \dots \subset N$ has simple or zero | |
quotients. | |
\end{lemma} | |
\begin{proof} | |
Indeed, for each $i$, $(N \cap M_i)/(N \cap M_{i-1})$ is a submodule of | |
$M_i / M_{i-1}$, so is either zero or simple. | |
\end{proof} | |
\begin{theorem}[Jordan-H\"older]\label{lengthexists} Let $M$ be a module of | |
finite length. | |
In this case, any two filtrations | |
on $M$ with simple quotients have the same length. | |
\end{theorem} | |
\begin{definition} | |
This number is called the \textbf{length} of $M$ and is denoted $\ell(M)$. | |
\end{definition} | |
\begin{proof}[Proof of \rref{lengthexists}] | |
Let us introduce a temporary definition: $l(M)$ is the length of the | |
\emph{minimal} filtration on $M$. We will show that any filtration of $M$ (with | |
simple quotients) is of length | |
$l(M)$. This is the proposition in another form. | |
The proof of this claim is by induction on $l(M)$. Suppose we have a filtration | |
\[ 0 = M_0 \subset M_1 \subset \dots \subset M_n = M \] | |
with simple quotients. We would like to show that $n = l(M)$. By definition of | |
$l(M)$, there is another filtration | |
\[ 0 = N_0 \subset \dots \subset N_{l(M)} = M. \] | |
If $l(M) = 0,1$, then $M$ is zero or simple, which will necessarily imply that $n=0,1$ | |
respectively. So we can assume $l(M) \geq 2$. We can also assume that the | |
result is known for strictly smaller submodules of $M$. | |
There are two cases: | |
\begin{enumerate} | |
\item $M_{n-1} = N_{l(M) -1 } $. Then $M_{n-1} = N_{l(M)-1}$ has $l$ at most | |
$l(M)-1$. Thus by the inductive hypothesis any two filtrations on $M_{n-1}$ | |
have the same length, so $n-1 = l(M) -1$, implying what we want. | |
\item We have $M_{n-1} \cap N_{l(M) - 1} \subsetneq M_{n-1}, N_{l(M)-1}$. | |
Call this intersection $K$. | |
Now we have two filtrations of these modules $M_{n-1}, N_{l(M)-1}$ whose | |
quotients are simple. We can replace them such that the next | |
term before them is $K$. | |
To do this, consider the filtrations | |
\[ 0 = M_0 \cap K \subset M_1 \subset K \subset \dots M_{n-1} \cap K = K | |
\subset M_{n-1} \] | |
and | |
\[ 0 = N_0 \cap K \subset M_1 \subset K \subset \dots N_{l(M)-1} \cap K = K | |
\subset N_{l(M)-1} . \] | |
These filtrations have simple or zero quotients by | |
\rref{simplefiltrationint}, and since $ M_{n-1}/K = | |
M_{n-1}/M_{n-1} \cap N_{l(M)-1} = M/M_{n-1}$ is simple, and similarly for | |
$N_{l(M)-1}/K$. We can throw out redundancies to eliminate | |
the zero terms. | |
So we get two new filtrations of $M_{n-1}$ and $N_{l(M)-1}$ whose second-to-last | |
term is $K$. | |
By the | |
inductive hypothesis any two filtrations on either of these proper submodules $M_{n-1}, | |
N_{l(M)-1} $ | |
have the same length. | |
Thus the lengths of the two new filtrations are $n-1$ and $l(M)-1$, | |
respectively. | |
So we find that $n-1 = l(K) +1$ and $l(M)-1 = l(K)+1$ by | |
the inductive hypothesis. This implies what we want. | |
\end{enumerate} | |
\end{proof} | |
\begin{exercise} | |
Prove that the successive quotients $M_i/M_{i-1}$ are also determined (up to | |
permutation). | |
\end{exercise} | |