Representations of Compact Groups (Part 1)
I've written a previous post on representation theory for finite groups. The representation theory of finite groups is very nice, but many of the groups whose representations we care about are not finite. For example, representations of are important for understanding the behavior of particles with nonzero spin. So we want to extend representation theory to more general groups. A nice family of groups to consider are compact groups. In many ways, compactness is a generalization of finiteness. To use an example from that link, every real-valued function on a finite set is bounded and attains its maximum. This is untrue for real-valued functions on infinite sets: consider the functions and respectively on the interval . However, continuous real-valued functions on a compact interval must be bounded and must attain their maxima. Similarly, compact groups generalize finite groups, and many of the nice features of the representation theory of finite groups extend to the representation theory of compact groups. This post will mostly follow the notes about representations of compact groups available here
Compact Topological Groups
First, we will start with some nice properties of compact topological groups. Recall that a topological group is a group endowed with a topology so that multiplication and the inverse map are continuous. When studying the representation theory of finite groups, it was often convenient to sum over the elements of the group (e.g. to define our inner product on the space of characters). Clearly we cannot always sum over the elements of an infinite group. But for compact groups, we have a nice theorem that tells us that we can integrate over the group instead, which is just as good.
This theorem is tricky to prove for locally compact topological groups. But for Lie groups, it is fairly easy. So we will just show a version of the theorem for Lie groups.
First, we will show existence. Let denote the dimension of . Recall that as long as is oriented, an -form on induces a measure. Furthermore, we recall that if we can find a nonvanishing -form on , then must be orientable. So it is sufficient to find a left-invariant nonvanishing -form on . Let denote the space of -covectors on the tangent space to the identity of . Pick any nonzero . Now, we can extend to a differential form on . Let denote the automorphism of given by left-multiplication by . This is continuous. sends to , so we can pull back along this map to define . This defines a differential -form on all of . is left-invariant by construction. Clearly this differential form is nonvanishing. And by negating if necessary, we see that is positive with respect to 's orientation, so it defines a left-invariant measure on .
Now, you might be wondering why it is important that is compact, because the above theorems don't require compactness. The nice thing about compactness is that measures only let us integrate functions with compact support - but if is compact, then every function has compact support. So we can integrate any real- (or complex-) valued functions on . We will write the integral of with respect to the Haar measure as .
In particular, we can integrate the constant function over compact groups. It is convenient to normalize our Haar measure so that . I will assume that all Haar measures are normalized in this way.
From now on, I'll assume that all groups are compact Lie groups unless I explicitly state otherwise.
Basic Definitions
We'll start with a whole bunch of definitions. They're essentially the same as the analogous definitions for finite groups, except we require that our maps are continuous. To do so, we have to put topologies on the vector spaces involved.
Useful Constructions
There are several simple subrepresentations we can consider.
- For any representation , is a subrepresentation because .
- Similarly, is a subrepresentation of itself.
- We also have a subrepresentation , the subspace of -invariants. Note that the action of on is trivial.
- Given any -linear map , the kernel is a subrepresentation of and the image is a subrepresentation of .
Given two representations and , there are several ways we can build new representations out of them.
- We can define a representation of on the dual space (where is the base field) by setting for .
- We can define a representation of on the conjugate space . We define as follows: it is the same topological abelian group as , but the scalar multiplication is changed. Let denote an element of and denote the corresponding element of . Then we set . That is to say, we scalar multiply by the conjugate of instead of by itself. The action of on is the same as the action of on .
- We can define a representation of on by setting .
- We can define a representation of on by setting .
- We can define a representation of on by using the isomorphism for finite-dimensional and using our constructions for taking tensor products and duals of representations.
Note that is a scalar and is a vector in . So this is just . Since acts by a linear map, we can factor out the to obtain So given any , we have .
First, suppose that . Then , so in particular we have for any . Using the formula for , we see that for all . Multiplying both sides by , we find that . Since this is true for all , we conclude that is -linear. So .
Conversely, suppose that . Then for all . So . Letting , we see that for all . So is in the subspace of invariants .
Complete Reducibility and Schur's Lemma
These all have determinant one, and are thus invertible. Furthermore, the product of two upper-triangular matrices is an upper-triangular matrix, so this is a group. This group has a natural action on given by the usual matrix-vector product. This defines a representation of on .
Note that this representation fixes the subspace given by
But it doesn't fix any other nontrivial subspaces. So is neither an irreducible representation nor a completely reducible representation of .
It's kind of frustrating that not all representations are completely reducible. One of the nice features of finite groups is that all representations of finite groups are completely reducible. We will show that compact groups are nice in this way as well- all representations of compact groups are completely reducible as well.
- Let be irreps of . Let . Then is either 0 or an isomorphism.
- Let be a complex irrep of . Then (i.e. any -linear endomorphism of is a scalar multiple of the identity)
- Since is -linear, we know that are subrepresentations. Since are irreps, this implies that is either or all of , and is either or all of . Thus, the only way for to be nonzero is if and . This means that if is nonzero, it must be an isomorphism.
- Since is a complex matrix, it has an eigenvalue . Clearly is a -linear endomorphism of . Thus, . But cannot be an isomorphism. So it must be . Thus, .
First, we will construct one such projection. Explicitly, we define This operation averages over the group action, which is why we named the projection . To show that is a projection, we have to show that it restricts to the identity on its image. First, we note that the image of is simply . We see that is contained in . Let be any vector in . Then for any , we have the last equality follows from the left-invariance of our measure. So we see that , which implies that .
Furthermore, any vector of is itself fixed by . If , then . So in particular, . Thus, we see that , and acts as the identity on its image. So it is a projection.
Now, we will show uniqueness. First, note that commutes with any other . Suppose that is another projection onto . In particular, it is an element of , so it commutes with . Thus,
Let be any hermitian inner product on . We can view as an element of . We can think of the -invariant inner products as elements of . So gives us a -invariant inner product.
Explicitly, this just means that we can define a -invariant inner product by the formula
Let be a -invariant inner product on . Let .
We note that is a subrepresentation of . Let . By definition, for all . Since the inner product is -invariant, . Since is a subreresentation, , so . Thus, for all , so we conclude that .
Therefore, .
Just reply Maschke's theorem repeatedly. Since our vector space is finite-dimensional, this process must terminate.
This follows from the above corollary and Shur's lemma.
Characters
Let be a (-linear) isomorphism. Then . So . By the cyclic property of the trace, . Thus,
- where
- where
- where considered as a constant function
- Since acts on by , and transposing does not change the trace, we see that .
- Since scalar multiplication on is conjugated, we have to take the complex conjugate of the entries in the matrix to get the matrix which acts on . Thus, .
- .
- .
- .
- This one is more complicated. We need to compute . To do so, we use a trick involving the averaging projection.
Note that the averaging projection acts as the identity on and acts as on the orthogonal complement to . Thus, acts as on and acts as on the orthogonal complement to . So . (Here denotes the trace over and denotes the trace over )
Therefore, Since our measure is left-invariant, this is just
The previous propositions tell us that characters of are a decategorification of the category of finite-dimensional representations of . Decategorification is the process of taking a category, identifying isomorphic objects and forgetting all other morphisms. This eliminates a lot of useful information, but often makes the category easier to work with. For example, if we decategorify the category of finite sets, we identify all sets with the same cardinality, and forget about all other functions. This just leaves us with the natural numbers, because sets are classified by their cardinality.
Frequently, there are nice structures in the category that still make sense after decategorification. For example, decategorifying disjoint unions of finite sets gives us addition of natural numbers, and decategorifying cartesian products gives us multiplication of natural numbers.
Above, we saw that characters are a decategorification of finite-dimensional -representations. Two characters are equal if the corresponding representations are isomorpic, and the direct sums, tensor products, etc. of -representations translate nicely into operations on characters. One of the most interesting aspects of this decategorification is that s turn into inner products.
Recall that a pair of adjoint functors are functors such that for all . Adjoint functors are so named in analogy with adjoint linear operators (Recall that two operators on a Hilbert space are adjoint if for all vectors .) This connection between inner products and Hom sets can be formalized to give a categorification of Hilbert spaces.
For an introduction to (de)categorification, you can look here (for a simpler introduction) or here (for a more complicated introduction).
Application: Irreducible Representations of
Before proceeding, let's use some of this machinery we have built up so far to find all irreducible representations of .
It will be helpful to use another characterization of as well.
Recall that the quaternions are defined by where . Then . We have a natural action of on by left-multiplication. This gives us a two-dimensional complex representation of . Writing it out explicitly, we see that Thus, our representation is given by This matrix is unitary, and has determinant . So this is clearly a continuous bijection from to . You can check that this bijection is a group isomorphism.
Since acts on , we also get an action of on , the space of complex polynomials in 2 variables. Given , we define We note that this action does not change the degree of monomials. Thus, the space of homogeneous polynomials of degree is invariant under this action. So it is a subrepresentation. Let denote the space of homogeneous polynomials of degree . We will show that are nonisomorphic irreducible representations, and every irreducible representation of is isomorphic to some . First, we'll start with a lemma about the structure of .
- Every element of can be written
- For fixed , is a 2D sphere with radius intersecting at
- Using our identification of with , we can think of points on the sphere as special unitary matrices. Unitary matrices are unitarily-diagonalizable. Clearly we can rescale these matrices so that the matrices we diagonalize with are in . Finally, note that a diagonal matrix in must have the form for , . Thus, diagonal matrices in correspond to points on the sphere.
-
Since the quaternion norm is multiplicative, and elements of have norm 1, we see that . Furthermore, Now, we will consider the map . Note that for unit quaternions, , the conjugate of . So we can also write this map . Note that So is purely imaginary. Furthermore, since and are all unit quaternions, so is their product. Thus, we can think of as a map , where we view as the unit quaternions and as the unit imaginary quaternions.
Furthermore, is surjective. If we represent vectors in as imaginary quaternions, then is a representation of on which acts by rotations. Since we can write all rotations in this form, and the rotation group acts transitively on the two-sphere, we see that covers all of . So is surjective.
So since , we see that is a sphere with radius . Now, we just have to check that the intersection of this sphere with is . Note that is the imaginary unit quaterions, and the only imaginary unit quaternions that lie in are . Thus, the intersection of with is .
To prove this, we will use characters. For convenience, let us write as . Recall that the image of in is the matrix Note that the eigenspaces of this operator on are with eigenvalues . Therefore, Note that all of these characters are different. This shows that all of the representations are distinct. Now, we will show that the characters are orthonormal.
Recall that the inner product on characters is given by Since the volume of is , we can write where is the standard volume element on . So we want to compute Recall that characters are constant on conjugacy classes. Since every element of is conjugate to exactly two unit complex numbers, we have Above, we showed that these orbits are spheres with radius . Therefore, the volume of an orbit is . Substituting this and our expressions for the characters, we see that our inner product is Because sines with different frequencies are orthogonal, we conclude that So our characters are orthonormal.
Finally, we will show that these are all of the irreducible representations. Suppose that was another irreducible representation. Then Using the same computational tricks, we see that Since sinces form an orthonormal basis for the set of square-integrable functions on the circle, we see that , which is impossible. Thus, every irreducible representation must be isomorphic to some .
Characters made it fairly easy to classify all of the irreducible representations of . Later on, we will generalize some of the computational techniques we used here to find the Weyl character formula and Weyl Integration Formula, which will be very useful for understanding representations. But that will have to be another post, since this one is already much longer than I realized it would be.
No comments: