Reductive groups: a rapid introduction

The goal of this post is to introduce, in a very informal way, the notion of a reductive group, and discuss some examples.

Preamble

Last term I gave a mini-course in a seminar here at Berkeley on the theory of linear algebraic groups with a focus on reductive groups. As part of that, I wrote up a significant set of notes (now in excess of 200 pages), which I have been wanting to finish and post. Of course, this is large undertaking, and so might take a while.

But, I just recently also gave a talk in a number theory seminar at Berkeley on essentially the same topic. The main difference was that this was a seminar on Shimura varieties (so mostly interested in the characteristic 0 theory), and I only had an hour and a half (opposed to 6 hour and a half lectures). Consequently, many of the proofs, but more importantly beautiful details of the theory, were left off.

While this is certainly not ideal for someone wanting to get a very thorough, and deeply-rooted understanding of linear algebraic groups, I realized that it serves an equally important purpose. Sometimes one just wants to get an idea of what something is, to see what it’s used for, and what the main tools used in the subject look like.

Thus, for this reason, I’ve decided to essentially transcribe my notes into wordpress format, with some slight expansions. I do still hope, eventually, to post my long set of notes from the mini-course, but hopefully, as I mentioned above, this post will serve as a rough guide, or, at least, an amuse-bouche.

The main goal is just to understand the definition of a reductive group, and understand how one would show various basic examples of linear algebraic groups are, in fact, reductive.

Motivation

The topic of this post is reductive linear algebraic groups. Let us motivate our reason to care about such objects adjective chunk by adjective chunk.

Groups

There are many fancy reasons one can give as to why the groups in a given category are desirable objects of study. But, to be frank, they mostly boil down to one simple idea: groups are nice.

Namely, it’s a founding principle of geometry that objects enjoying a large sets of symmetries are generally much nicer behaved, and have a much richer theory, than those without such transformations. Groups then, in this light, are perhaps the most obvious source of objects with large sets of symmetries since any group G acts simply transitively on itself by left translation.

Of course, really what we’re talking about are homogenous spaces, spaces with a transitive action by their automorphism group, or equivalently, quotients of groups. But, since groups themselves are the simplest type of such spaces, they are of initial interest.

Thus, an algebraic geometer should be interested in  algebraic groups for the simple reason that they provide, in some sense, the simplest type of geometric object. They are also the objects for which an abstract notion of symmetry can be studied algebraically. Less cryptically, if one has a flat projective T-scheme X, then it’s automorphism functor S\mapsto \text{Aut}(X_S) is representable, and thus a group scheme.

These are some of the reasons, amongst many others, that we care about groups.

Linear algebraic

But, the topic of this post is not about all algebraic groups but, instead, about linear algebraic groups. So, why do we make this restriction? There are two reasons that immediately come to mind.

First, linear algebraic groups are those that are accessible by representation theory. And since, after all, representation theory is one of the most powerful and comfortable tools available to mathematicians, this singles out linear algebraic groups as being a natural first subcategory of all groups to consider.

The second reason is that, for all intents and purposes, not much is lost by just studying algebraic groups. This is codified by the following famous result of Chevalley:

Theorem(Chevalley): Let k be a perfect field, and G/k a finite type group scheme. Then, there is a (essentially unique) decomposition:

1\to H\to G\to A\to 1

in the category of group schemes, where H is a linear algebraic group, and A is an abelian variety.

Thus, in some sense, the entire theory of group schemes is partitioned into two subcases—the theory of linear algebraic groups, and the theory of abelian varieties.

Unfortunately, or very fortunately, these two subjects, while both extremely rich, are largely incommensurate. Namely, what essentially characterizes linear algebraic groups is that they have representation theory (in a sense made precise below) whereas every representation of a connected abelian variety is trivial.

But, even if one is entirely interested in the abelian variety side of this picture (which some number theory enthusiasts might believe) the study of such objects is greatly enhanced by the study of linear algebraic groups. One reason, to be remarked on in the next section, is via the theory of Shimura varieties (which, very roughly, are moduli spaces of abelian varieties).

Another, more direct reason, is via the theory of Néron models. Namely, an abelian variety A/\mathbb{Q}_p need not admit an abelian scheme model \mathscr{A}/\mathbb{Z}_p. But, the very general theory of Néron models provides a smooth group scheme \mathscr{G}/\mathbb{Z}_p with generic fiber A subject to some natural conditions.

Many of the properties of A can be understood well in terms of the geometry and algebra of this group scheme \mathscr{G} and its special fiber \mathscr{G}_{\mathbb{F}_p}. And, thanks to the theorem of Chevalley mentioned above, this is always broken into an abelian variety component and a linear algebraic group component. Thus, even if one wants to understand abelian varieties, one is often times forced to deal with linear algebraic groups.

As a concrete example of this, recall the following famous theorem of Néron-Ogg-Shafarevich (or, perhaps more appropriately, Ogg-Tate-Serre):

Theorem(Néron-Ogg-Shafarevich): Let R be a DVR with perfect residue field, K=\text{Frac}(R), and A/K an abelian variety. Then, A has good reduction (i.e. admits an abelian scheme model \mathscr{A}/R) if and only if the usual Galois representation

\rho:G_K\to\text{GL}_{\mathbb{Z}_\ell}(T_\ell(A))

is unramified for any \ell invertible on R.

This theorem is extremely important. Practically it comes up in many other parts of one’s study of abelian varieties over number fields (e.g. Shafarevich’s conjecture, the open image theorem, etc.). Philosophically, it says that the (étale) homology of an abelian A variety is sufficiently rich enough to know the reduction type of A—a very deep idea.

I bring this theorem up here since the proof of said theorem breaks roughly into two parts. First is an understanding of Néron models. The second is a general understanding of the structure of linear algebraic groups. Once one has those two things the proof is, in essence, ‘easy’.

Reductive

We come, finally, to the adjective ‘reductive’ a, at first, scary sounding concept.

Reductivity can be thought about in a very simple-minded way as the property of a linear algebraic group G the guarantees that G has ‘nice representation theory’. This is a bit of an oversimplification, but serves one well intuitively.

For the purposes of the study of Shimura varieties reductiveness has a much more concrete realization. Namely, once again only roughly, Shimura varieties can be thought about as moduli spaces of polarizable Hodge structures. Then reductive groups become an obvious part of the study since they are, in essence, the Mumford-Tate groups of polarizable Hodge structures.

If the second bit of motivation means nothing to you, and it might not, then latch onto the first since, historically, this was the motivation for the study of reductive groups. Especially since this niceness of representation theory makes reductive groups amenable to ‘classification’.

Basics

Basic results

Before we walk, we crawl. In our case here, that means that we should recall the very basic ideas and definitions in the theory of group schemes.

In particular, let’s start with the fundamental definition. Let us fix a field k. We then say that a group scheme over k consists of a scheme G/k together with maps

m:G\times_k G\to G

and

e:\text{Spec}(k)\to G

and

i:G\to G

such that the quadruple (G,m,i,e) ‘is a group’. In fancier parlance, this means that group schemes are group objects in the category of schemes.

It is often times much more fruitful to think about group schemes through a functor of points lens. Namely, by Yoneda’s lemma, to give G/k the structure of a group scheme (i.e. the maps m, i, and e) it is equivalent to give a functorial in T group structure on G(T) for every k-scheme T.

In other words, for every scheme G/k one has the contravariant Hom functor

\text{Hom}_k(-,G)=:h_G:(\mathsf{Sch}/k)^\text{op}\to\mathsf{Set}

and to give G the structure of a group is to factorize h_G through the forgetful functor \mathsf{Grp}\to\mathsf{Set}. Thus, a group scheme over k is a functor F:(\mathsf{Sch}/k)^\text{op}\to\mathsf{Grp} such that the composition of F with the forgetful functor \mathsf{Grp}\to\mathsf{Set} is representable.

Thus, in practice, and surely in this post, we will often times think about group schemes as just being group-valued contravariant functors on \mathsf{Sch}/k. We will, without comment, take for granted that these functors are, in fact, representable (when thought about set-valued functors) leaving the details of this to the reader.

Let us now give some examples of group schemes over k:

  • Fix a finite abelian group A. Consider the functor which takes a k-scheme T to the group A^{\pi_0(T)} (where \pi_0(T) is the number of connected components of T). This is a group scheme over k called the constant groups scheme with values in A.
  • Let \mu_n be the group scheme over k with functor of points given by \mu_n(T)=\{\alpha\in\mathcal{O}_T(T):\alpha^n=1\}. This is called the group scheme of n^\text{th}-roots of unity.
  • Let \text{GL}_n be the group scheme over k with functor of points given by T\mapsto \text{GL}_n(\mathcal{O}_T(T)). This is called the general linear group. It is almost universally true that one denotes \text{GL}_1 by \mathbf{G}_m and calls it the multiplicative group.
  • Let \text{SL}_n be the group scheme which assigns to a k-scheme T the group \text{SL}_n(\mathcal{O}_T(T)). This is called the special linear group.
  • Let \langle-,-\rangle denote the standard orthogonal pairing on A^n for any ring A. Then, we can consider the groups \text{O}_n and \text{SO}_n which assign to a scheme T those elements of \text{GL}_n(\mathcal{O}_T(T)) which preserve this pairing in the case of \text{O}_n, and those which preserve the pairing and have determinant 1 in the case of \text{SO}_n. We call these the orthogonal group and special orthogonal group respectively.
  • Now let \langle-,-\rangle denote the standard symplectic pairing on A^{2n} for any ring A. We define two group schemes \text{Sp}_{2n} and \text{GSp}_{2n} which associates to a T the subgroup of \text{GL}_{2n}(\mathcal{O}_T(T)) which preserves the symplectic pairing in the case of \text{Sp}_{2n}, and those which preserve the pairing up to a (fixed) unit scalar in the case of \text{GSp}_{2n}. We call these the symplectic group and symplecitic similitude group.
  • We define the group scheme \mathbf{G}_a^n to associate to any T the group \mathcal{O}_T(T)^n.

Just as an example, in the above, \mu_n is represented by \text{Spec}(k[T]/(T^n-1)) and the general linear group \text{GL}_n by \text{Spec}(k[x_{ij},(\det)^{-1}]).

As one might expect, a morphism of group schemes is just a morphism which preserves the multiplication map or, equivalently, a natural transformation between their associated group functors.

So, let us now define our actual term of interest. We call a group scheme G/klinear algebraic group if G/k is finite type and affine. The term is not a misnomer since:

Theorem 1: Let G/k be a group scheme. Then, G/k is a linear algebraic group if and only if there exists an embedding G\hookrightarrow\text{GL}_n for some n.

Here the embedding means a closed embedding of groups which is just a group map whose underlying map of schemes is a closed embedding.

Thus Theorem 1 tells us that linear algebraic groups are just closed subgroups of the general linear groups—they are groups with faithful representations. While the proof of Theorem 1 is not too difficult, there is something semi-subtle happening. Indeed, one can define a group scheme over any scheme T, and the definition of a linear algebraic group still stands—finite type and affine. But, the analogy of Theorem 1 is open if one replaces k by even something as simple as k[\varepsilon]:=k[x]/(x^2) (although it’s known for DVRs).

Now, one could spend literally years trying to understand the geometry of a general group scheme, or even just of a linear algebraic group (cf. SGA 3), but that is not our goal here. Thus, let us suffice ourselves with two theorems concerning such geometry.

The first goes to show how the homogenous space property makes the study of groups much simpler:

Theorem 2: Let G/k be a linear algebraic group. Then, G is smooth if and only if G is geometrically reduced. Thus, if k is perfect, then G is smooth if and only if it’s reduced. If k has characteristic 0, then G/k is automatically smooth.

One nice corollary of the above is the following. There are many theorems below which, personally, I find very exciting. They all say something of the form ‘the only class of groups which satisfy BLAH conditions are the obvious ones’. Now, usually this class BLAH of conditions is really natural except for the inclusion of ‘smooth’ which, often, seems like an afterthought. The above tells us that as long as we’re in characteristic 0 we can completely ignore this condition—all groups are smooth!

Proof(Sketch): Obviously if G is smooth it’s geometrically reduced. Conversely, if G/k is geometrically reduced then G_{\overline{k}}/\overline{k} is reduced. But, then it’s well known (cf. Tag 056V) that there is a dense open U\subseteq G_{\overline{k}} for which U is regular. But, then for any other closed point g\in G(\overline{k}) one sees that gU\subseteq G_{\overline{k}} is a smooth neighborhood of g, and thus g is regular. Thus, all the closed points of G_{\overline{k}} are regular, thus G_{\overline{k}} is regular, thus G is smooth.

The second claim is a famous theorem of Chevalley.\blacksquare

To summarize this proof: smoothness happens on a dense open for reduced schemes, but groups are homogenous (every point has a neighborhood isomorphic to a neighborhood of every point) so this smoothness propagates over the whole scheme. This line of argument is very common.

The other geometric idea we’ll need is the notion of exact sequence, quotients, and the existence thereof. Now, this is one of the most complicated parts of the basic theory of group schemes—what does G/H mean? This is, almost everywhere in scheme theory, a difficult question—quotients are hard.

In this context, we have the following definition though. Given a map f:G\to H of group schemes over k the kernel \ker f is the naive kernel (i.e. it’s G\times_H \{e_G\}) and \text{coker}(f) is the fppf sheafification of the quotient presheaf T\mapsto G(T)/H(T).

The shocking, stupendous theorem is then the following:

Theorem 3: Let G be a linear algebraic group over k, and H\subseteq G a closed subgroup. Then, G/H exists. If H is normal in G then this quotient has the unique structure of a group scheme such that the canonical map G\to G/H is a group map.

This is surprisingly complicated result (see SGA 3), but one we’ll use freely through the rest of this post without much comment.

Let us explore the above point just a tiny bit since, frankly, it can be quite confusing. In particular, let us focus our attention on what might be the most confusing application of Theorem 3. Namely, let us consider two group schemes over k (assumed to be characteristic 0 for convenience). Namely:

\text{PSL}_n:=\text{SL}_n/\mu_n\qquad\qquad\text{PGL}_n:=\text{GL}_n/Z(\text{GL}_n)

where, of course, this quotient is taken in the sense of Theorem 3.

Now, let us begin by making the following surprising observation: \text{PSL}_n=\text{PGL}_n. Namely, there is a canonical map \text{PSL}_n\to\text{PGL}_n and, in fact, it’s an isomorphism. Indeed, it’s a closed embedding of group schemes obviously, but it’s also a surjection since this can be checked on \overline{k}-points where its obvious.

While this is shocking, it’s made less so if we recall what \text{PSL}_n really is. Namely, for a second, let us denote the group scheme \text{PSL}_n as H_n. Then, the first important thing to notice is that H(T)\ne \text{PSL}_n(\mathcal{O}_T(T)) where, on the right, this is the ‘usual group’ \text{SL}_n(\mathcal{O}_T(T))/\mu_n(\mathcal{O}_T(T)). Why is this true? Well, it’s because \text{SL}_n(\mathcal{O}_T(T))/\mu_n(\mathcal{O}_T(T)) is the value on T of the quotient presheaf of \text{SL}_n/\mu_n but H_n is the quotient sheaf (in the fppf topology).

In fact, to cement this point, note that the association

T\mapsto \text{SL}_n(\mathcal{O}_T(T))/\mu_n(\mathcal{O}_T(T))

is not even representable! In particular, it is not a sheaf, so that we really do have to sheafify (for which it’s then still not obvious its representable, which is the content of Theorem 3).

Indeed, let’s see why this is true when k=\mathbb{Q} and n=2. For notational convenience (i.e. not to confuse us) let’s denote the association T\mapsto \text{SL}_n(\mathcal{O}_T(T))/\mu_n(\mathcal{O}_T(T)) by P_n(T). Then, we’re trying to show that P_2 is not a sheaf for the fppf topology. If P_2 were an fppf sheaf, in particular an étale sheaf, then we’d have for any prime p that

P_2(\mathbb{Q})\xrightarrow{\approx} P_2(\mathbb{Q}(\sqrt{p}))^{G}\subseteq P_2(\mathbb{Q}(\sqrt{p})

would be a bijection where G:=\text{Gal}(\mathbb{Q}(\sqrt{p})/\mathbb{Q}) acts in the usual way.

That said, note that

\displaystyle \begin{pmatrix}\sqrt{p} & 0\\ 0 & \displaystyle \frac{1}{\sqrt{p}}\end{pmatrix}

is in P_2(\mathbb{Q}(\sqrt{p}))^G. Indeed, the non-trivial element \sigma of G acts on this matrix by

\displaystyle \sigma:\begin{pmatrix}\sqrt{p} & 0\\ 0 & \displaystyle \frac{1}{\sqrt{p}}\end{pmatrix}\mapsto \begin{pmatrix}-\sqrt{p} & 0\\ 0 & \displaystyle \frac{-1}{\sqrt{p}}\end{pmatrix}

but these two elements of P_2(\mathbb{Q}(\sqrt{p})) are equal since \mu_2(\mathbb{Q}(\sqrt{p}))=\{\pm 1\}. That said, this matrix is not in the image of P_2(\mathbb{Q}). Thus, we see that P_2 is not a sheaf, and so, by fiat, not representable.

So, the question remains—what are the sections H_n(T) if not P_n(T)? The answer, as one might guess, lies in étale cohomology. Namely, we have a short exact sequence of fppf sheaves

1\to\mu_n\to\text{SL}_n\to H_n\to 1

and thus for every T we get an exact sequence

1\to \mu_n(T)\to\text{SL}_n(T)\to H(T)\to H^1_\text{fppf}(T,\mu_n)

and thus an exacts sequence

1\to P_n(T)\to H_n(T)\to H^1_\text{fppf}(T,\mu_n)

for all T.

Now, since we assumed that we’re working in characteristic 0 we have that

H^1_\text{fppf}(T,\mu_n)=H^1_{\acute{e}\text{t}}(T,\mu_n)

and from standard Kummer theory we know that we have a short exact sequence

0\to \mathcal{O}_T(T)^\times/(\mathcal{O}_T(T)^\times)^n\to H^1_{\acute{e}\text{t}}(T,\mu_n)\to\mathrm{Pic}(T)[n]\to 0

Thus, we see that the natural map P_n(T)\to H_n(T) is an isomorphism whenever \mathcal{O}_T(T)^\times=(\mathcal{O}_T(T)^\times)^n and \mathrm{Pic}(T)[n]=0. So, for example, we see that

H_n(\mathbb{P}^1_{\overline{\mathbb{Q}}})=P_n(\mathbb{P}^1_{\overline{\mathbb{Q}}})=\text{SL}_n(\overline{\mathbb{Q}})/\mu_n(\overline{\mathbb{Q}})

for all n.

It’s perhaps easier to understand the H_n(T) is using the fact that H_n is isomorphic to the algebraic group \text{PGL}_n which, for reasons similar to the above, we’ll denote G_n (again, recall though, that G_n=H_n we denote them differently here purely for cosmetic reasons). Namely, calling it G_n is wise since G_n(T)\ne \text{GL}_n(\mathcal{O}_T(T))/\mathcal{O}_T(T)^\times (where this denominator is embedded as invertible scalar matrices). Let us denote this latter quantity \text{GL}_n(\mathcal{O}_T(T))/\mathcal{O}(T)^\times by P'_n(T)—again, it’s the quotient presheaf \text{GL}_n/Z(\text{GL}_n).

Again though, we have an exact sequence for all T

1\to P'_n(T)\to G_n(T)\to H^1_{\acute{e}\text{t}}(T,\mathbf{G}_m)

But, now, we know that

H^1_{\acute{e}\text{t}}(T,\mathbf{G}_m)=\mathrm{Pic}(T)

and thus we see that

H_n(T)=G_n(T)=P'_n(T)=\text{GL}_n(\mathcal{O}_T(T))/\mathcal{O}_T(T)^\times

whenever \mathrm{Pic}(T)=0.

Thus, as an example of this, we see that

H_n(\mathbb{Q})=\text{GL}_n(\mathbb{Q})/\mathbb{Q}^\times

even though H_n is the group scheme \text{PSL}_n!

Of course, we will stop denoting \text{PSL}_n by H_n and \text{PGL}_n by G_n—we’ll just denote them as \text{PSL}_n and \text{PGL}_n. That said, the above shows how subtle these notions are and how confusing the naming of them is. So, one should not sweat these details too much except to be very mindful that most group schemes like \text{PSL}_n only necessarily give the correct group on geometric points (on other points one has to make a cohomology computation).

Applications to elliptic curves

To end this section, I’d like to give one way in which even these basic ideas allow us to better understand number theoretic questions.

To begin with, let us say that a group scheme T/k is a torus if T_{\overline{k}} (or equivalently T_{k^\text{sep}}) is isomorphic to a power of the multiplicative group. Let us give some examples of these objects.

  • Let \mathbb{S} denote the following subgroup of \text{GL}_{2,\mathbb{R}}:

    \displaystyle \mathbb{S}:=\left\{\begin{pmatrix}a & b\\ -b & a\end{pmatrix}:a^2+b^2\ne 0\right\}

    Then, one can show that this is a 2-dimensional non-split (i.e. not just a power of the multiplicative group rationally) torus. It’s called the Deligne torus. One might recognize this construction since, in essence, this is just the standard embedding of \mathbb{C}^\times into \text{GL}_2(\mathbb{R}) done on the level of schemes.

  • Consider the following subgroup of \mathbb{S}

    \mathrm{Res}^1_{\mathbb{C}/\mathbb{R}}\mathbf{G}_{m,\mathbb{C}}:=\left\{\begin{pmatrix}a & b\\ -b & a\end{pmatrix}:a^2+b^2=1\right\}\subseteq\text{GL}_{2,\mathbb{R}}

    then one can show fairly easily that this is a non-split 1-dimensional torus over \mathbb{R}. In fact, there is a pretty natural short exact sequence

    1\to\mathrm{Res}^1_{\mathbb{C}/\mathbb{R}}\mathbf{G}_{m,\mathbb{C}}\to \mathbb{S}\to\mathbf{G}_{m,\mathbb{R}}\to 1

So, how do tori help us understand number theory better? Well, the key is the following two theorems which, at first, seem intimidating but are really not:

Theorem 4: Let k be a field. Then, there is a canonical bijection between the isomorphism classes of 1-dimensional tori over k and the set \text{Hom}_\text{cont.}(G_k,\mathbb{Z}/2\mathbb{Z}) where \mathbb{Z}/2\mathbb{Z} has the discrete topology.

Explicitly, this says that there is a bijection between 1-dimensional tori and the fields k and the quadratic Galois extensions L/k. What is this bijection? Well, k goes to the split torus \mathbf{G}_{m,k} and L/k goes to \text{Res}^1_{L/k}\mathbf{G}_{m,L}. Understanding this precisely won’t be important to our application at large.

But, one can make an even more powerful statement which, to me, is a truly spectacular, beautiful result that one would not expect should exist:

Theorem 5: Let k be a perfect field and G/k a 1-dimensional connected smooth linear algebraic group. Then, G is one of the following:

  1. \mathbf{G}_a.
  2. \mathbf{G}_m
  3. \text{Res}^1_{L/k}\mathbf{G}_{m,L} for some Galois quadratic extension L/k.

In particular, we deduce the following:

Corollary 6: Let \mathbb{F}_q be, as usual, a finite field. Then, up to isomorphism, there are only three 1-dimensional smooth connected linear algebraic groups over k. They are \mathbf{G}_{a,\mathbb{F}_q}, \mathbf{G}_{m,\mathbb{F}_q}, and \text{Res}^1_{\mathbb{F}_{q^2}/\mathbb{F}_q}\mathbf{G}_{m,\mathbb{F}_{q^2}}.

OK, this is nice, but so what? I’ve still not explained how this helps us understand something number theoretic!

So, finally, let us consider what is the beginning of many number theoretic jokes: so an elliptic curve over \mathbb{Q} with bad reduction at `17 walks into a bar…

But, really, let E/\mathbb{Q} be an elliptic curve. Then, since \mathbb{Q} has class number 1 we can lift E to a curve \mathcal{E}/\mathbb{Z} in a canonical way (its minimal Weierstrass model). We then define the reduction of E at p to be \mathcal{E}_{\mathbb{F}_p}.

Now, it’s well known that \mathcal{E}_{\mathbb{F}_p}^\text{sm}, the smooth subset of \mathcal{E}_{\mathbb{F}_p}, has the structure of a group (the same chord-tangent construction works). So, since \mathcal{E}_{\mathbb{F}_p}^\text{sm} is a smooth, connected, 1-dimensional group scheme over \mathbb{F}_p, and all curves are either affine or projective, we know that \mathcal{E}_{\mathbb{F}_p}^\text{sm} is either a smooth, connected, 1-dimensional linear algebraic group, or it’s still an elliptic curve.

Thus, by Corollary 6 there are four choices for what \mathcal{E}^\text{sm}_{\mathbb{F}_p} is. These four choices have names:

\mathcal{E}_{\mathbb{F}_p}^\text{sm}=\begin{cases}\text{An elliptic curve} & \mbox{if}\quad p\text{ is a prime of good reduction}\\ \mathbf{G}_{a,\mathbb{F}_p} & \mbox{if}\quad p\text{ is a prime of bad, additive reduction}\\ \mathbf{G}_{m,\mathbb{F}_p} & \mbox{if}\quad p\text{ is a prime of bad, split multiplicative reduction}\\ \mathrm{Res}^1_{\mathbb{F}_{p^2}/\mathbb{F}_p}\mathbf{G}_{m,\mathbb{F}_{p^2}} & \mbox{if}\quad p\text{ is a prime of bad, non-spilt multiplicative reduction}\end{cases}

so the (at least to me) mysterious ‘four reduction types’ can be understood quite nicely in terms of the group schemes over a finite field.

Reductive group schemes

We now move onto the main course, the theory of reductive group schemes.

Before we begin in earnest, let us say, again, what the intuition is: reductive group schemes are those with nice representation theory. This manifests itself in characteristic zero in the following nice way:

Theorem 7: Let k be characteristic 0. Then, G/k a linear algebraic group is reductive if and only if \mathsf{Rep}(G) is semisimple.

Here \mathsf{Rep}(G) is the category of representations (i.e. homomorphisms into some general linear group). Thus, in words, the above says that reductive groups are precisely those groups for which every representation breaks up into irreducible parts. This is clearly a highly desirable property of a group and is the reason for the namesake ‘reductive’ (things reduce to irreducible pieces!).

Of course, in positive characteristic things get a little more sticky. There the notion of reductive and linearly reductive (i.e. all your representations are semisimple) diverge (the latter implies the former, but not conversely). But, for technical reasons, it’s the notion of reductiveness which stays more true to our ideals especially when concerned with classification theory.

For example since reductive groups are supposed to be ‘nice groups’ we’d hope that the most natural of all groups, the general linear group, is reductive. This is true, but in characteristic p the group \text{GL}_p need not be linearly reductive. Thus, again for technical reasons, reductiveness and not linear reductiveness, is the correct notion even though, personally, I mostly think about reductive as meaning linearly reductive up to some fudge factor.

So, if linearly reductive is not the right notion we need to find some other way of defining reductiveness. The rough idea is to single out those groups with the worst representation theory, and define reductive groups as being those groups with no such part.

To this end, we make the following definition. A linear algebraic group U/k is called unipotent if it satisfies any of the following equivalent conditions:

  • It admits an embedding into

    U_n:=\left\{(a_{ij})\in\text{GL}_n:a_{ii}=1\text{ for all }i,\text{ and }a_{ij}=0\text{ if }i>j\right\}

  • All representations have a fixed point. In other words, U has no non-trivial simple representation.

  • It admits a decomposition

    \{1\}=U_0\subseteq U_1\subseteq\cdots \subseteq U_n=U

    with each U_i normal in U_{i+1} and U_{i+1}/U_i is isomorphic to a subgroup of \mathbf{G}_a.

  • If U is also smooth, then there is the following equivalent condition. There exists an embedding U\hookrightarrow \text{GL}_n such that U(\overline{k}) (or, more properly, the image of this group) consists entirely of unipotent matrices (i.e. a matrix A such that A-I is nilpotent).

It should be noted that in this third criteria, if U/k is smooth and connected (which is automatic if k has characteristic 0) one can take U_{i+1}/U_i\cong\mathbf{G}_a.

Let us give some simple examples:

  • The group U_n is obviously unipotent.
  • The group \mathbf{G}_a^n is evidently unipotent by the third criteria.
  • Assume that k has characteristic p. Let \alpha_p be the group scheme given by T\mapsto \{\alpha\in\mathcal{O}_T(T):\alpha^p=0\}. One can think of \alpha_p as being the kernel of the Frobenius map F:\mathbf{G}_{a,k}\to\mathbf{G}_{a,k} given on T points by x\mapsto x^p. This is, of course, a group map because we’re in characteristic p. Then, \alpha_p is unipotent since it’s a subgroup of \mathbf{G}_a.

We can see the second criteria as saying that U/k has the worst possible representation theory. Namely, the only semisimple representation is the trivial representation! Thus, one might imagine that those G/k which have good representation theory are those with no unipotent part.

To this end, let us define the unipotent radical of a linear algebraic group G, denoted R_u(G), to be the largest connected, smooth, normal, unipotent subgroups of G.

The first non-obvious fact is the following:

Theorem 8: R_u(G) exists.

Let us give some examples of unipotent radicals:

  • R_u(T)=\{1\} for any torus T.
  • R_u(\text{GL}_n)=\{1\}.
  • R_u(\text{SL}_n)=\{1\}.
  • R_u(\text{Sp}_{2n})=\{1\}.
  • R_u(U_n)=U_n.
  • Let B_n denote the subgroup of \text{GL}_n of all upper triangular matrices. Then, R_u(B_n)=U_n.
  • R_u(\mathbf{G}_a^n)=\mathbf{G}_a^n.

Let us also state one other basic property of the unipotent radical:

Theorem 9: For any separable extension L/k one has that R_u(G)_L=R_u(G_L).

Thus, we might imagine that groups with nice representation theory are those with R_u(G)=0. So, to this end, we can now give a rigorous of a reductive group. Let G/k be a linear algebraic group. Then, G is reductive if R_u(G_{\overline{k}})=\{1\}.

Of course, the first thing that one might ask is whether we can remove this geometric condition—whether G being reductive is equivalent to R_u(G)=\{1\}. If k is perfect, so that \overline{k}=k^\text{sep}, then this follows immediately from Theorem 9. If not, then this needn’t be the case. Lucky for you, dear reader, Brian Conrad et. al. has written an entire book about the topic called Pseudo-reductive Groups. Feel free to check it out! 🙂

So, we now have a bunch of examples of reductive groups:

  • \text{GL}_n.
  • \text{SL}_n.
  • \text{Sp}_{2n}.
  • \text{GSp}_{2n}.
  • \text{O}_n.
  • \text{SO}_n.
  • Tori.

and some good non-examples

  • Any unipotent group.
  • B_n.

But, these claims are all rested upon the computations claimed above (and some that weren’t claimed), such as R_u(\text{GL}_n)=\{1\}. Thus, we’d like to work towards understanding how such a computation can be made in the rest of this post.

But, before we move on to this, we’d like to discuss some basic properties of reductive groups and how one can understand them very well in the case when k=\mathbb{R}.

To begin, let us state the properties of reductive groups, in the abstract, we’d like to emphasize:

Theorem 10: Let G/k be a linear algebraic group.

  1. If k has characteristic 0, then G is reductive if and only if it’s linearly reductive.
  2. G is a torus if and only if G is connected, smooth, reductive, and commutative.
  3. The group G/R_u(G) is reductive. Thus, every linear algebraic group is the extension of a reductive group by a unipotent group.

We’ve already discussed 1. of the above theorem. Property 3. says that, up to the extension problem. all linear algebraic groups can be understood in terms of reductive and unipotent groups. Both type of groups are relatively simple—reductive groups have nice representation theory (and, in fact, a classification theorem!), and unipotent groups are just iterated extensions of \mathbf{G}_a.

We’d like to then just mention why 2. is true by discussing a more general fact:

Theorem 11: Let G/k be a connected smooth commutative linear algebraic group. Then, G\cong U\times_k T where U is unipotent and T is a torus.

Thus, 2. of Theorem 10 follows quite easily. Indeed, in such a decomposition G\cong U\times_k T it’s evident that R_u(G)=U. Thus, R_u(G)=\{1\} (i.e. G is reductive) if and only if G=T, so that G is a torus.

So, as stated above, let us finish this section with a look at what reductiveness looks like when k=\mathbb{R}.

The key definition is the following. Let G/\mathbb{R} be a linear algebraic group. Then, an involutive morphism \theta:G\to G is called a Cartan involution if

G^{\theta}(\mathbb{R})=\left\{g\in G(\mathbb{C}):\theta(\overline{g})=g\right\}

which is a Lie group, is compact. One can, in fact, show that, as the notion suggests, there is an algebraic group G^\theta for which G^\theta(\mathbb{R}) is just the \mathbb{R}-points.

The main theorem is then the following:

Theorem 12: Let G/\mathbb{R} be a linear algebraic group.

  1. All Cartan involutions on G are conjugate.
  2. G is reductive if and only if it has a Cartan involution.

Proof(Sketch): Let us say why one direction of 2. is true. Namely, suppose that G has a Cartan involution. To show that G is reductive it suffices, by Theorem 9, to show that G_\mathbb{C} is reductive. Since we’re in characteristic 0 it suffices to show that G_\mathbb{C} is linearly reductive.

Let V be any representation of G_\mathbb{C}. Let \theta be a Cartan involution and let G^\theta be the twist by \theta. Note that we obtain a representation of the Lie group G^\theta(\mathbb{R}) which, since it’s a compact Lie group, is semisimple. Since the \mathbb{R}-points are Zariski dense, one can use the decomposition of this Lie group representation to produce a decomposition of the G_\mathbb{C}-representation. \blacksquare

Let us give what is, essentially, the only example of a Cartan involution:

  • Consider the involution \theta of \text{GL}_{n,\mathbb{R}} given by \theta(A)=(A^{-1})^\top—inverse transpose. This is a Cartan involution since those A\in\text{GL}_n(\mathbb{C}) such that \theta(A)=\overline{A} are precisely the unitary matrices which, as is well-known, form a compact subgroup of \text{GL}_n(\mathbb{C}).

Using the above ideas, we derive the following extremely pleasing result:

Theorem 13: Let G/\mathbb{R} be a linear algebraic group. Then, G is reductive if and only if there exists a faithful representation G\hookrightarrow \text{GL}_n whose image is stable under transpose.

Proof(Sketch): If such an embedding exists, then the restriction of \theta to G would give a Cartan involution of G. The other direction is more complicated and requires some legwork of Satake. \blacksquare

This immediately proves that all the claimed examples of reductive groups above are reductive, at least in the case when k=\mathbb{R}. But, one can show that for a linear algebraic group G/\mathbb{Q} reductiveness is equivalent to reductiveness of G_\mathbb{R}/\mathbb{R}. Thus, we immediately conclude that all of the following groups are reductive over \mathbb{Q} and thus reductive over any characteristic 0 field: \text{GL}_n, \text{SL}_n, \text{Sp}_{2n}, \mu_n, \text{GSp}_{2n}, \text{O}_n, …

But, frankly, this feels like cheating. Namely, even though we have rigorously proven reductiveness we have done it by recourse to analytic techniques—the fact that all compact Lie groups are linearly reductive requires the existence of G-invariant inner product, which, in turn, requires using the Haar measure (i.e. ‘average’ any inner product).

Thus, we’d still like to justify our calculations of R_u(G) as above using purely algebraic techniques, which is what we will aim to do.

Semisimple groups

Our goal of trying to understand computations such as R_u(\text{GL}_n)=\{1\} so as to able to, in particular, give a characteristic independent way of proving \text{GL}_n is reductive, is going to naturally take us through the world of semisimple groups.

One can intuitively think about semisimple groups as being reductive groups with finite center or, equivalently, reductive groups which are almost adjoint (which, as in the theory of Lie groups, means that G\to\text{GL}(\text{Lie}(G)) is injective). We will explain at the end of the post why ‘semisimple’ is not a misnomer since it won’t at all be clear from our initial presentation of such groups.

So, semisimple groups are going to come from forbidding large normal subgroups of a certain sort. But, instead of forbidding large normal unipotent subgroups we’re going to forbid large normal solvable subgroups.

So, we begin, as we should, with the definition of solvable. Namely, let G/k be a smooth linear algebraic group. Then, we say that G is solvable if either of the equivalent properties holds:

  • The group G(\overline{k}) is solvable.
  • There is a filtration

    \{1\}=G_0\subseteq G_1\subseteq\cdots\subseteq G_n=G

    such that G_i is normal in G_{i+1} and G_{i+1}/G_i is abelian.

One must be careful since these two are only equivalent when G is smooth.

Let us give some good examples:

  • Any unipotent group.
  • The group B_n—it has U_n as as normal subgroup with quotient \mathbf{G}_m^n.

In fact, the second group above is fairly representative as the following shows:

Theorem 14 (Lie-Kolchin Theorem): Let G/\overline{k} be a smooth solvable connected group. Then, G admits an embedding G\hookrightarrow B_n.

There are also some representative non-examples:

  • \text{GL}_n.
  • \text{SL}_n.

In fact, it’s not hard to show that

[\text{GL}_n(\overline{k}):\text{GL}_n(\overline{k})]=\text{SL}_n(\overline{k})

[\text{SL}_n(\overline{k}):\text{SL}_n(\overline{k})]=\text{SL}_n(\overline{k})

which certainly shows the result. Once one understands the slightly subtle notion of an algebro-geometric definition of commutator subgroup (which we discuss later in this post), the above can be made algebraic.

So, as stated above, semisimple groups are made by essentially restricting them from having large normal solvable subgroups. So, we need an analogue of R_u(G) which is supplied by the following. Let G be a linear algebraic group. Define the radical of G, denoted R(G), to be the largest normal, smooth, connected, solvable subgroups of G.

Again, the first non-obvious theorem is the following:

Theorem 15: Let G/k be a linear algebraic group. Then:

  1. R(G) exists.
  2. R(G)_L=R(G_L) for any separable extension L/k

Let us consider some examples:

  • R(\text{GL}_n)=Z(\text{GL}_n), in other words its the scalar matrices.
  • R(\text{SL}_n)=\{1\}.
  • R(\text{Sp}_{2n})=\{1\}.
  • R(G)=G for any connected smooth solvable group.

Again, it’s not obvious how to actually make these computations—something we will remedy below.

So, with this setup we can easily make the definition of semisimple groups. Namely, a linear algebraic group G/k is semisimple if R(G_{\overline{k}})=\{1\}. Similar comments concerning equivalence to R(G)=\{1\} hold as in the case of reductive groups. Namely, as long as k is perfect this is equivalent to R(G)=\{1\} and in the non-perfect case one gets, what I assume, would be called pseudo-semisimple groups.

Some examples:

  • \text{SL}_n.
  • \text{Sp}_{2n}.
  • \text{PSL}_n which is defined as \text{SL}_n/\mu_n.

and some non-examples:

  • Solvable groups.
  • \text{GL}_n.
  • \text{GSp}_{2n}

Now, we want to develop some machinery that will allow us to actually make these computations. The first step in this chain is to discuss the notion of a Borel subgroup. Namely, a closed subgroup B\subseteq G is Borel if it is connected, smooth, solvable and B_{\overline{k}} is maximal amongst subgroups of G_{\overline{k}} with this property.

The reason for needing maximality over \overline{k} is very much related to the following: Borel’s need not exist. Namely, a linear algebraic group G/k which has a Borel is called quasi-split. It’s not hard to show that if k=\overline{k} then all groups are quasi-split but if k is, say, \mathbb{Q} then there are certainly non-quasi-split groups.

Checking that a group is Borel, and classifying all such Borels seems like a difficult problem. And, while it is, the following certainly makes it much more tenable:

Theorem 16: Let G/k be a linear algebraic group.

  1. All Borels are conjugate over \overline{k}.
  2. A connected solvable smooth subgroup N\subseteq G is Borel if and only if G/N is projective (equivalently, proper).

Thus, for example, we can now give some good examples of Borels:

  • The subgroup B_n\subseteq\text{GL}_n is Borel. Indeed, it’s evidently smooth, connected, and solvable. Thus, it remains to show that \text{GL}_n/B_n is projective. But, \text{GL}_n acts transitively on the complete flag variety \mathscr{F}\ell with stabilizer of the standard flag being B_n. Thus, \text{GL}_n/B_n\cong\mathscr{F}\ell. And, since \mathscr{F}\ell is projective, it follows that B_n is a Borel.

The reason that Borels are powerful in the computation of radicals and unipotent radicals is the following:

Theorem 17: Let G/k be a linear algebraic group where k=\overline{k}. Then, the following equality holds:

\displaystyle R(G)=\left(\bigcap_{B\text{ Borel}}B\right)^\circ_\text{red}

Here the \circ denotes connected component, and the subscript \text{red} means take the reduced subscheme (which is still a group).

Proof(Sketch): Since R(G) is connected, solvable, and smooth it’s contained in a maximal such group which is necessarily Borel. But, since all Borels are conjugate, and R(G) is normal, R(G) is contained in every Borel so contained in their intersection. It’s then contained in the connected component since it’s connected, and contained in the reduced subscheme since it’s smooth.

Conversely, the intersection of all Borels is normal (since they’re all conjugate) and solvable (since this can be checked on \overline{k}-points). Thus, the intersection of the Borels connected component’s reduced structure is connected, normal, solvable, and smooth and thus, by definition, contained in R(G). \blacksquare

We need one last tiny, obvious observation:

Theorem 18: Let G/k be a linear algebraic group. Then:

\displaystyle R_u(G)=R_u(R(G))

Thus, we are finally prepared to justify some of the calculations we made above:

  • We claim that R(\text{GL}_n)=Z(\text{GL}_n) and R_u(\text{GL}_n)=\{1\}. It suffices to check this geometrically. To see this, begin by noting that, as proved above, B_n is a Borel subgroup of \text{GL}_n. Thus, also, is B_n^{-}, the group of lower triangular matrices. Thus, we see that R(G)\subseteq B_n\cap B_n^{-} which consists of diagonal matrices. Then, one can easily see that the only normal-in-\text{GL}_n subgroup of the diagonal matrices is Z(\text{GL}_n) which proves the claim.Note then that by Theorem 18 R_u(\text{GL}_n) is R_u(Z(\text{GL}_n)). But, evidently, R_u(Z(\text{GL}_n))=\{1\}.
  • We claim that R(\text{SL}_n)=R_u(\text{SL}_n)=\{1\}. It suffices to check this geometrically, and to only check that R(\text{SL}_n)=\{1\}. But, again, note that if we define B=B_n\cap\text{SL}_n and B^{-}=B_n^{-}\cap\text{SL}_n that R(\text{SL}_n)\subseteq B\cap B^{-}. But, B\cap B^{-} is just the diagonal matrices of determinant 1. The only normal-in-\text{SL}_n subgroup of this is just the diagonally embedded copy of \mu_n. But, the largest connected reduced subgroup of \mu_n is \{1\} and since R(\text{SL}_n) must be contained in this subgroup, the conclusion follows.

One can use similar ideas to do the rest of the computations that we claimed above. In particular, one can use this to prove the groups we claimed are reductive/semisimple really are so (independent of characteristic).

Let us end this section with some general calculation friendly results:

Theorem 19: Let G be a connected reductive linear algebraic group over k where k is perfect. Then:

R(G)=Z(G)^\circ_\text{red}

Here the subscript ‘\mathrm{red}‘ means the reduced subscheme which, because k is perfect, inherits the structure of a group.

Proof(Sketch): We obviously have that Z(G)^\circ_\text{red}\subseteq R(G), so it suffices to check the reverse inclusion.

Begin by noticing that R(G) is a torus since G is reductive. Indeed, by Theorem 18 it’s clear that R(G) is reductive. But, then R(G) is a connected, reductive, solvable group which implies it’s a torus. Indeed, since R(G) is connected and solvable it has a subnormal filtration whose quotients are abelian, reductive, connected groups. But, by Theorem 11 this implies that R(G) has a subnormal filtration whose quotients are tori. One can show this implies that R(G) is a torus.

Now, the next key step is the following rigidity lemma:

Lemma 20: Let G/k be a connected linear algebraic group and T/k a torus. Then, any action of G on T is trivial.

From this, we conclude that the conjugation action of G on the torus R(G) is trivial, and so R(G) is central. Thus, R(G)\subseteq Z(G)^\circ_\text{red} giving the reverse inclusion. \blacksquare.

With this, we can justify our intuition from the beginning of the section:

Theorem 21: Let G/k be a connected linear algebraic group where k is perfect. Then, G is semisimple if and only if G is reductive and Z(G) is finite (as a k-scheme).

Proof(Sketch): Suppose first that G is semisimple. Then, since Z(G)^\circ_\text{red}\subseteq R(G) we deduce that Z(G)^\circ_\text{red}=\{1\}. This evidently implies that Z(G) is zero-dimensional, and so finite.

Conversely, suppose that G is reductive with finite center. Then, Z(G) being finite means that Z(G)^\circ_\text{red}=\{1\} and so the conclusion follows from Theorem 19. \blacksquare

These theorems also allow us to try and understand semisimple and reductive groups in terms of their commutative subgroups. Namely:

Theorem 22: Let G/k be a connected linear algebraic group and assume that k is perfect. Then, G is reductive if and only if the largest normal abelian subgroup is of multiplicative type. It’s semsimple if and only if its only normal abelian subgroups are finite étale.

Here, to be of multiplicative type means that, geometrically, one is isomorphic to product of tori and groups of the form \mu_n for n coprime to \mathrm{char}(k).

Proof: Suppose that G is reductive and let H be any normal commutative subgroup. Then, H^\circ is normal connected and commutative, and so H^\circ\subseteq R(G). As was shown in the proof of Theorem 19, we know that R(G) is a torus. Then, H^\circ is a torus by fiat.

Now, as classically is known (see section 5.i of Milne’s notes on algebraic groups) there is a short exact sequence

1\to H^\circ \to H\to \pi_0(H)\to 1

where \pi_0(H) is a finite étale scheme (the group of components). Now, since k is perfect this sequence splits (see 5.53 of loc. cit.) and since our groups are abelian we conclude that

H\cong H^\circ\times \pi_0(H)

But, \pi_0(H) is a finite abelian étale group scheme, and so multiplicative. The result follows.

Conversely, suppose that every normal abelian subgroup of G is of multiplicative type. Then, in particular, R(G) is of multiplicative type and, since it’s connected, must be a torus. This then implies that

R_u(G)=R_u(R(G))=0

as desired.

Suppose now that G is semisimple and let H\subseteq G be any normal abelian subgroup. Then, H is of multiplicative type (since G is reductive). But, since H^\circ\subseteq R(G)=0 we conclude that H is finite, as desired.

Conversely, suppose that G has only finite étale subgroups. Then, note that G is reductive (by the previous case) and so by Theorem 19 R(G)=Z(G)^\circ. But, since Z(G) is a normal commutative subgroup, this implies that Z(G) is finite abelian étale. This implies that R(G)=Z(G)^\circ is trivial (this can be checked over \overline{k} where all finite abelian étale groups are constant). \blacksquare

This gives a fairly concrete way of discussing semisimplicity and reductivity of groups since, as is plain to see, it doesn’t require any of the more sophisticated notions of unipotent groups, solvable groups, etc. This is the perspective taken in notes such as this nice article..

The isogeny category

Alternative characterizations

In this section we’d like to discuss, without proof (or without the entire proof), some structure results about reductive and semisimple groups.

To begin we’d like to approach the question of finding ‘nice groups’ in a different way. Above, we essentially approached this question from the point of view that ‘nice groups’ meant groups with reasonably nice representation theory. But, there is another obvious perspective. Namely, in a subject like group theory we define an object to be ‘simple’ if it has no non-trivial subobjects, and this is one good definition of ‘nice’. Can something similar happen here? If so, how do reductive and semisimple groups fit into that picture?

Let us begin with some definitions. A group G is called simple if it is non-abelian and has no non-trivial, proper, normal subgroups. The exclusion of abelian is largely because they ‘play differently’ than the other simple groups (in the same way that \mathbb{Z}/p\mathbb{Z} is the odd man out in the theory of finite simple groups).

Her are some examples:

  • The group \text {PSL}_n=\text{PGL}_n is simple for n>1–this is not overly obvious. In characteristic 0 it suffices to check that \text{PSL}_n(\overline{k}) is simple in which case one can see this nice note of K. Conrad.
  • The group \text{SL}_n is almost simple. Indeed, if N\unlhd \text{SL}_n is a non-trivial proper connected smooth subgroup, then the same would true of the image in \text{PGL}_n which implies, by the previous bullet, that N\subseteq \mu_n. Thus, every non-trivial proper normal subgroup of \text{SL}_n is finite (in fact, contained in \mu_n!)
  • The group \text{GL}_n is not simple.

Also, let us call a homomorphism of algebraic groups G\to G' an isogeny if it is finite and surjective with kernel lying in Z(G). We say that two groups G and G' are isogenous if there exists a third group J and isogenies J\to G' and J\to G.

Remark: What we have defined here is really called a central isogeny. The difference is moot if G\to G' is smooth (e.g. if k is characteristic 0) but can differ in positive characteristic. This is not important for us here though.

Also, note that the inclusion of this J in the above can be somewhat disturbing for those familiar with abelian varieties. Namely, if there is an isogeny of abelian varieties A\to B, then there is an isogeny of abelian varieties B\to A. Thus, the notion of ‘isogenous’ for abelian varieties is just the existence of an isogeny in one direction which, by the above statement, really is an equivalence relation. That said, while there is an isogeny \text{SL}_n\to\text{PGL}_n there is no isogeny \text{PGL}_n\to\text{SL}_n (because \text{SL}_n is simply connected and \text{PGL}_n\not\cong\text{SL}_N) and thus the same definition as for abelian varieties wouldn’t work here.

Let us give an example

  • The obvious projection map \text{SL}_n\to\text{PSL}_n is surjective with kernel \mu_n, and so an isogeny.
  • The map \text{SL}_n\times \mathbf{G}_m\to\text{GL}_n sending, on T-points, the pair (A,\alpha) to \alpha A is an isogeny with kernel isomorphic to \mu_n (specifically the kernel is pairs of the form (\alpha A,\alpha^{-1}) where \alpha is an n^\text{th}-root of unity). This example is indicative of the ‘general case’ of reductive groups.

Thus, with these definitions we can give the following pleasant rephrasing of reductiveness and semisimplicity in the isogeny category:

Theorem 23: Let G/k be a linear algebraic group. Then:

  1. G is reductive if and only if G is isogenous to a product of simple groups and tori.
  2. G is semisimple if and only if G is isogenous to a product of simple groups.

This justifies the name ‘semisimple’ and also explains that, at least in the isogeny category, reductive and semisimple groups are precisely the ‘other type’ of ‘nice group’ one gets by decomposing groups into their ‘simplest pieces’.

We can’t hope to prove 2. in Theorem 23  here (it’s incredibly lengthy) but, if we assume it, we can give a fairly simple proof of 1.

To start, let us define a notion which is deceptively difficult. Namely, for a linear algebraic group G/k we’d like to define the derived subgroup (also called the commutator subgroup) D(G). Now, the naive guess is that we should define it as follows:

D(G)(T):=D(G(T))

but this is wrong! For example, we would very much like that for any field k we have the equality

D(\text{GL}_n)=\text{SL}_n

Indeed, this is what happens on \overline{k}-points and so, of course, what should actually happen. But, for example one has that

D(\text{GL}_2(\mathbb{F}_2))\ne \text{SL}_2(\mathbb{F}_2)

This example, while silly, does highlight the fact that our definition of the derived subgroup needs to be slightly more thought out than just defining it ‘pointwise’.

So, with this in mind, we give the alternate definition. Namely, let us define the derived subgroup D(G) of a linear algebraic group G/k to be the intersection of all normal algebraic subgroups N of G such that G/N is abelian. Intersection here can seem a little dodgy at first glance, but it’s actually A-OK. Namely, take the scheme theoretic intersection of these N in G, and note that it’s still a subgroup scheme. Be careful though, don’t be fooled into thinking that we can just take the set-theoretic intersection of the N and give it the smallest closed subscheme structure (i.e. the reduced structure). While this works swimmingly in characteristic 0 noting that in \text{GL}_p the intersection of the smooth subgroups Z(\text{GL}_p) and \text{SL}_p which gives \mu_p shows that things are slightly more complicated in positive characteristic.

One can put their minds at ease concerning this definition if G is smooth and connected (which essentially all of our examples are). Specifically, one then has that D(G) is characterized as the unique smooth connected subgroup H such that H(k^\text{sep})=D(G(k^\text{sep})) (and similarly with k^\text{sep} replaced by \overline{k}).

Some examples of the above are as follows:

  • D(\text{GL}_n)=\text{SL}_n.
  • D(\text{SL}_n)=\text{SL}_n.
  • D(G)=0 for any abelian linear algebraic group G.
  • D(\text{GSp}_{2n})=\text{Sp}_{2n}.
  • D(B_n)=U_n.

And, as is plain to see, if G is connected and smooth, then G is solvable (in the sense used above) if and only if the derived chain

G\supseteq D(G)\supseteq D(D(G))\supseteq\cdots

terminates in finite time. Indeed, since all groups in sight are smooth, it suffices to check this on \overline{k}-points. But, the fact that D(G)(\overline{k})=D(G(\overline{k})) then completes the proof.

So, how does this help us with Theorem 23? Well, we have the following observation:

Theorem 24: Let G be a connected smooth reductive group over k, then D(G) is semisimple.

The key to proving Theorem 24 will be the following beautiful result:

Theorem 25: let G be a connected smooth reductive group over k where k is perfect. Then, R(G)\cap D(G) is finite, and R(G)D(G)=G.

Here R(G)D(G) means the smallest algebraic subgroup of G containing R(G) and D(G).

Proof(Sketch): To see that R(G)\cap D(G) is finite, we may proceed as follows. It suffices to check finiteness over \overline{k}, so let’s assume that k is algebraically closed. Note then that for any faithful representation \rho:G\hookrightarrow \text{GL}(V) we may, since R(G) is a torus, conjugate this embedding so that \rho(R(G))(k) lands in the diagonal matrices of \text{GL}(V). In fact, we can decompose V into to character spaces for the R(G)-action, say we decompose it as V=V_1\oplus\cdots\oplus V_n with V_i acting by the character \chi_i. In particular, for any g\in R(G)(k) we have that \rho(g) is of the form \text{diag}(M_1,\ldots,M_n) where M_i is a scalar matrix (specifically with scalar \chi_i(g)).

Note though that since R(G)\subseteq Z(G) we must have that \rho(G(k)) lands in the centralizer of the matrices of the form \text{diag}(M_1,\ldots,M_n) with M_i scalar, which is precisely the group \text{GL}(V_1)\times\cdots\times\text{GL}(V_n) embedded diagonally into \text{GL}(V). So, finally note that if g\in (R(G)\cap D(G))(k) then \rho(g) is both of the form \text{diag}(M_1,\ldots,M_n) and a commutator in \text{GL}(V_1)\times\cdots\times\text{GL}(V_n) which is equivalent to being in \text{SL}(V_1)\times\cdots\times \text{SL}(V_n). But, the matrices M_i are then both scalar and have determinant 1, which means that their entries are \dim(V_i)-roots of unity, of which there are only finitely many. Thus, (D(G)\cap R(G))(k) is finite as desired.

To see the second claim consider the quotient G/(R(G)D(G)) this is possible since both R(G) and D(G) are normal (and thus so is the group generated by the two). Note that this group is semisimple since it’s a quotient of G/R(G) (which is semisimple) and also abelian (being a quotient of G/G^\text{der}). But, since it’s also connected, the claim follows from Theorem 19. \blacksquare

We can now prove Theorem 24:

Proof(Theorem 24, Sketch): Note that the map D(G)\to G/R(G) is finite surjective. But, G/R(G) is semisimple. So, suppose that N\unlhd D(G) is connected smooth and solvable. Then, evidently, its image in G/R(G) is connected normal smooth solvable and thus, since G/R(G) is semisimple must be trivial. This implies that N\subseteq D(G)\cap R(G) and thus finite. But, since N is smooth and connected this is a contradiction. \blacksquare

Thus, we can finally explain how 1. of Theorem 23 follows from 2. Indeed, consider the map

R(G)\times D(G)\to G

given, on T-points, by (g,h)\mapsto gh. Note that this is, in fact, a group map since (again by Theorem 19) we know that R(G)\subseteq Z(G). This map is surjective by Theorem 25 and has kernel R(G)\cap D(G) which, again by Theorem 25, is finite and evidently central. Thus, G is isogenous to R(G)\times D(G). But, D(G) is semisimple (by theorem 24) and R(G) is a torus (by theorem 19). Thus, if we know 2. then certainly we can deduce 1.

For example, we already remarked that \text{GL}_n was isogenous to \text{SL}_n\times\mathbf{G}_m which is precisely the isogeny with D(\text{GL}_n)\times Z(\text{GL}_n). As we’ve already remarked \text{SL}_n is isogenous to the semisimple group \text{PGL}_n and \mathbf{G}_m is a torus, and thus a decomposition of \text{GL}_n (in the isogeny category) as a product of a simple group a torus is given by \text{PGL}_n\times\mathbf{G}_m.

The isogeny theorem

Let us end this post with just a taste of the very in-depth classification theory that accompanies reductive and semisimple groups. This is, broadly, the theory of root data which is an entire subject onto itself. So, we’ll suffice it to state here one version of the isogeny theorem.

To state this correctly though, we need one more definition. Namely, we call a connected semisimple groups G simply connected if every isogeny G'\to G, with G' connected, and whose kernel lies in Z(G') is an isomorphism. In characteristic 0 this second condition is automatic. Indeed, the action of G on the finite kernel must be trivial since G is connected and the kernel is finite étale.

The name can be justified as follows:

Theorem 26: let G/\mathbb{C} be a connected semisimple group. Then, G is simply connected if and only if G^\text{an} is simply connected.

Here G^\text{an} denotes the analytification of G—the associated complex Lie group.

Proof(Sketch): Suppose first that G^\text{an} is simply connected. Note that any isogeny G'\to G induces, in particular, a covering map (G')^\text{an}\to G^\text{an}. Thus, by assumption, this is an isomorphism, so G'\to G is an isomorphism.

Conversely, suppose that G is simply connected in the sense of semisimple groups. We aim to show that \pi_1^\text{top}(G^\text{an})=0. But, since G^\text{an} is Lie group we know that \pi_1^\text{top}(G^\text{an}) is a finitely generated abelian group and, consequently, it suffices to show that its profinite completion is zero. But, it’s profinite completion is, as is well-known, \pi_1^{\acute{e}\text{t}}(G). Thus, it suffices to show that \pi_1^{\acute{e}\text{t}}(G)=0. But, if G'\to G is a Galois étale cover, then it’s not hard to show that one can supply G' the structure of  a semisimple group such that this map is an isogeny, so it must be an isomorphism. Thus, \pi_1^{\acute{e}\text{t}}(G)=0 as desired. \blacksquare

In fact, less obviously so (but in the same vein as the above theorem) is the following. Let k be a perfect field and and let G be a connected semisimple group. Then, there exists a a simply connected semisimple group G_{\text{sc}} and an isogeny G_{\text{sc}}\to G called the universal cover of G. It satisfies the property that for any other central isogeny H\to G, there is a unique factorization through G_{\text{sc}}\to G. If one then defines \pi_1^\text{alg}(G):=\ker(G_{\text{sc}}\to G), which can be shown to a finite connected commutative  group scheme, then one can show that if \mathrm{char}(k)=0 then \pi_1^\text{alg}(G) is non-canonically isomorphic to \pi_1^\text{geom.}(G) (where, recally, one defines \pi_1^\text{geom}(X):=\pi_1^{\acute{e}\text{t}}(X_{\overline{k}})).

So, with these definitions down, we can state one basic version of the isogeny theorem:

Theorem 27 (Isogeny theorem): Let k be algebraically closed, and G/k a simply connected semisimple group. Then, G is isogenous to a product of groups G_i where G_i is an element of the following list:

  1. \text{SL}_{n+1}
  2. \text{Sp}_{2n}.
  3. \text{Spin}(n).
  4. One of several ‘exceptional’ groups.

One might then wonder how to parameterize the isomorphism classes of semisimple groups within a fixed isogeny class. The idea, roughly, is that there is a ‘largest’ and a ‘smallest’ member of each isogeny class, and so really one just needs to understand what’s in between these two members. Specifically, recall that if G is any semisimple group then we have its universal cover G_{\text{sc}}\to G which is, by fiat, the largest group in G‘s isogeny class. The smallest is taken up by G^\text{ad} the adjoint group of G defined, as usual, to be G/Z(G). Moreover, the map G_{\text{sc}}\to G^\text{ad} is an isogeny, so we really only have to worry about finitely many such groups within a given isogeny class.

Note, moreover, that while the simply connected or adjoint groups in a given isogeny class are the obvious representatives, there really is no preferred representative between them. In particular, in Theorem 27 above list we gave the simply connected constituents of the isogeny classes of the groups involved. We could have also listed the adjoint constituents and said that G was isogenous to a product of groups in the following list:

  1. \text{PSL}_n
  2. \text{PSp}_{2n}
  3. \text{PSO}_n
  4. One of several ‘exceptional’ groups (which are now the adjoint exceptional groups).

which is really no different (and not wholly more enlightening).

12 comments

  1. Nice post as usual. Glad to see you’re back 🙂 P.S. It would be great if you finished these notes, and remember perfect is the enemy of good 🙂

  2. This post is incredibly helpful! Thank you!
    Just wanted to point out a possible minor typo: when you define \alpha_p, as an example of a unipotent group, you write p\alpha=0 … should it be \alpha^p=0?

  3. In the section on reduction of elliptic curves: why should the reduction of an elliptic curve mod p connected? I thought the number of connected components depends on p-adic valuation of the j-invariant…

    1. Hey Mayor,

      Are you thinking about the Neron model? I have not written down the Neron model, but the highly singular minimal Weierstrass models (the smooth locus of which is isomorphic to the connected component of the Neron model). I think you can look at the minimal Weierstrass model and see it has connected reduction.

      As a way overkill answer I think the following is correct. Suppose that X/\mathbb{Z} is integral and proper+flat(=surjective) over \text{Spec}(\mathbb{Z}), such that X_\mathbb{Q} is geometrically connected, then for all p\in\text{Spec}(\mathbb{Z}) one has that X_{\mathbb{F}_p} is connected. This should follow, essentially, from Zariski’s connectedness theorem.

      In our case we’re looking at \text{Proj}(\mathbb{Z}[X,Y,Z]/(f(X,Y,Z)) where f(X,Y,Z) is a (monic in Y^2Z) homogenous cubic. Everything I think is made right by the assumption that f(X,Y,Z) is irreducible in \mathbb{Q}[X,Y,Z]. So, the last paragraph applies (again this is overkill here).

      I apologize if I’ve made a silly mistake or misunderstood you!

      1. Oh, maybe I was being an idiot. Indeed, I was thinking of Néron model. Zariski’s main theorem doesn’t tell me if the smooth locus is connected.

        But now that I think about it, because Weierstrass model is minimal so the fibers are irreducible (for otherwise one could blow down components of fibers).

  4. Dear Alex,

    An exciting introduction! In the Néron-Ogg-Shafarevich theorem you stated, do you need perfectness of the residue field of $R$? Or could you give us a reference of that?

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s