Skip to main content

Section 1.3 \(\bmit{\SU(2)}\)

Subsection 1.3.1 Representations

Just as orthogonal groups preserve the magnitude of real vectors, the unitary groups preserve the magnitude of complex vectors. The “rotation” group in two complex dimensions is known as \(\SU(2)\text{.}\) If \(v\in\CC^2\text{,}\) its magnitude is defined by

\begin{equation} |v|^2 = v^\dagger v\tag{1.3.1} \end{equation}

where \(\dagger\) denotes the conjugate transpose, that is

\begin{equation} v^\dagger = \bar{v}^T .\tag{1.3.2} \end{equation}

The magnitude of a complex vector is still a non-negative real number, just as for real vectors.

So if we want to preserve the magnitude of \(v\text{,}\) we must have

\begin{equation} (Mv)^\dagger (Mv) = (v^\dagger M^\dagger)(M v) = v^\dagger v\tag{1.3.3} \end{equation}

or simply

\begin{equation} M^\dagger M = \one .\tag{1.3.4} \end{equation}

Matrices satisfying (1.3.4) are called unitary matrices; the “U” in \(\SU(2)\) stands for unitary. As before, the “S” stands for special, and refers to the additional condition that

\begin{equation} |M| = 1 .\tag{1.3.5} \end{equation}

Thus, we have

\begin{equation} \SU(2) = \{M\in\CC^{2\times2}:M^\dagger M=\one, |M|=1\} .\tag{1.3.6} \end{equation}

It's not that difficult to find the most general matrix satisfying (1.3.4) and (1.3.5), but let's first ask how many degrees of freedom there are. A \(2\times2\) complex matrix has 8 real degrees of freedom, the matrix equation (1.3.4) imposes four constraints, and the determinant condition (1.3.5) imposes one more constraint. Thus, we expect 3 parameters in our general element—just as for \(\SO(3)\text{.}\)

Let's start with some examples. Our old friend \(M(\phi)\in\SO(2)\) is in \(\SU(2)\text{,}\) but we give it a new name, writing

\begin{equation} S_y(\alpha) = \begin{pmatrix} \cos\alpha \amp -\sin\alpha \\ \sin\alpha \amp \cos\alpha \end{pmatrix} .\tag{1.3.7} \end{equation}

Other examples of unitary matrices are

\begin{equation} S_x(\alpha) = \begin{pmatrix} \cos\alpha \amp -i\sin\alpha \\ -i\sin\alpha \amp \cos\alpha \end{pmatrix} , \qquad S_z(\alpha) = \begin{pmatrix} e^{-i\alpha} \amp 0 \\ 0 \amp e^{i\alpha} \end{pmatrix} .\tag{1.3.8} \end{equation}

Subsection 1.3.2 Derivatives

We know what to do next: Differentiate! So consider

\begin{equation} \begin{aligned} s_y \amp= \dot{S}_y = S_y'(0) = \begin{pmatrix} 0 \amp -1 \\ 1 \amp 0 \end{pmatrix} ,\\ s_x \amp= \dot{S}_x = S_x'(0) = \begin{pmatrix} 0 \amp -i \\ -i \amp 0 \end{pmatrix} ,\\ \quad s_z \amp= \dot{S}_z = S_z'(0) = \begin{pmatrix} -i \amp 0 \\ 0 \amp i \end{pmatrix} .\end{aligned}\tag{1.3.9} \end{equation}

These matrices are linearly independent, and therefore span the tangent space at the identity.

So what? Well, let's compute the commutators. The result is

\begin{equation} [s_x,s_y] = 2s_z , \qquad [s_y,s_z] = 2s_x , \qquad [s_z,s_x] = 2s_y .\tag{1.3.10} \end{equation}

In other words, the commutators resulting from \(\SU(2)\) are the same  1  as those resulting from \(\SO(3)\text{.}\) So the infinitesimal versions of these two groups are the same. So they're the same group—at least locally.

Subsection 1.3.3 Comparison with \(\bmit{\SO(3)}\)

We can realize this identification explicitly by considering matrices of the form

\begin{equation} X = \begin{pmatrix}z\amp x-iy\\ x+iy\amp -z\end{pmatrix} .\tag{1.3.11} \end{equation}

The matrix \(X\) is tracefree, that is

\begin{equation} \tr(X) = 0 ,\tag{1.3.12} \end{equation}

and Hermitian, that is,

\begin{equation} X^\dagger = X .\tag{1.3.13} \end{equation}

We can act on \(X\) with \(\SU(2)\) via the action

\begin{equation} X \longmapsto MXM^\dagger\tag{1.3.14} \end{equation}

which preserves both of these conditions. Furthermore, we have  2 

\begin{equation} |X| = -(x^2+y^2+z^2) ,\tag{1.3.15} \end{equation}

and, since \(|M|=1\text{,}\) we have

\begin{equation} |MXM^\dagger| = |X| .\tag{1.3.16} \end{equation}

Thus, \(M\) can be identified with an element of \(\SO(3)\text{!}\)

However, that's not the whole story. Since there are two \(M\)s in our action, we can not tell the actions of \(\pm M\) apart. Thus, \(\SU(2)\) is the double cover of \(\SO(3)\text{.}\) These two groups have different global structure, but their local structure is the same.

Every orthogonal group \(\SO(n)\) admits such a double cover, but the double cover is not usually a unitary group. The double cover of \(\SO(n)\) is called \(\Spin(n)\text{;}\) we have shown that

\begin{equation} \Spin(3) \cong \SU(2) .\tag{1.3.17} \end{equation}

But our interest is in the local structure, and \(\SU(2)\) is the same as \(\SO(3)\) locally.

Subsection 1.3.4 Properties

You may recognize the \(s_m\) as being \(i\) times the corresponding Pauli matrices, that is,

\begin{equation} s_m = -i\sigma_m ,\tag{1.3.18} \end{equation}

so that

\begin{equation} \begin{aligned} \sigma_x \amp= \begin{pmatrix} 0 \amp 1 \\ 1 \amp 0 \end{pmatrix} ,\\ \sigma_y \amp= \begin{pmatrix} 0 \amp -i \\ i \amp 0 \end{pmatrix} ,\\ \sigma_z \amp= \begin{pmatrix} 1 \amp 0 \\ 0 \amp -1 \end{pmatrix} .\end{aligned}\tag{1.3.19} \end{equation}

What are the properties of the Pauli matrices? They are Hermitian, they are tracefree, and they square to the identity, that is

\begin{equation} \begin{aligned} \sigma_m^\dagger \amp= \sigma_m ,\\ \tr(\sigma_m) \amp= 0 ,\\ \sigma_m^2 \amp= \one .\end{aligned}\tag{1.3.20} \end{equation}

Similarly, the \(s_m\) are anti-Hermitian, tracefree, and square to minus the identity, that is

\begin{equation} \begin{aligned} s_m^\dagger \amp= -s_m ,\\ \tr(s_m) \amp= 0 ,\\ s_m^2 \amp= -\one .\end{aligned}\tag{1.3.21} \end{equation}

More generally, if \(M(\alpha)\) is any family of elements of \(\SU(2)\) such that \(M(0)=\one\text{,}\) then \(M(\alpha)^\dagger M(\alpha)=\one\) by (1.3.4). Differentiating this equation, we obtain

\begin{equation} M'(\alpha)^\dagger M(\alpha) + M(\alpha)^\dagger M'(\alpha) = 0\tag{1.3.22} \end{equation}

and evaluating this result at \(\alpha=0\) yields

\begin{equation} A^\dagger + A = 0\tag{1.3.23} \end{equation}

where we have written \(A\) for \(\dot{M}=M'(0)\text{.}\) This argument is true quite generally: If the elements of a Lie group are unitary matrices, then the elements of the corresponding Lie algebra are anti-Hermtian. Thus, we can describe the Lie algebra \(\su(2)\) of \(\SU(2)\) as

\begin{equation} \su(2) = \{A\in\CC^{2\times2}:A^\dagger+A=0, \tr(A)=0\} .\tag{1.3.24} \end{equation}

The same argument shows that infinitesimal orthogonal matrices are antisymmetric, as you can verify in the case of \(\SO(3)\) and \(\so(3)\text{.}\) Thus,

\begin{equation} \begin{aligned} \SO(3) \amp= \{M\in\RR^{3\times3}:M^T M=\one, |M|=1\} ,\\ \so(3) \amp= \{A\in\RR^{3\times3}:A^T+A=0, \tr(A)=0\}\end{aligned}\tag{1.3.25} \end{equation}

and we have shown that

\begin{equation} \su(2) \cong \so(3)\tag{1.3.26} \end{equation}

not merely as vector spaces, but as Lie algebras (since commutators are preserved).

The factor of \(2\) can be eliminated easily by dividing each of our basis elements \(s_m\) by \(2\text{,}\) or equivalently by replacing \(\alpha\) by \(\alpha/2\) in \(S_m\text{.}\)
The determinant \(\det(X)\) of a square matrix \(X\) is often written as \(|X|\text{,}\) a practice we adopt. Do you see why this usage makes sense?