# Peeter Joot's (OLD) Blog.

• ## Archives

 Adam C Scott on avoiding gdb signal noise… Ken on Scotiabank iTrade RESP …… Alan Ball on Oops. Fixing a drill hole in P… Peeter Joot's B… on Stokes theorem in Geometric… Exploring Stokes The… on Stokes theorem in Geometric…

• 294,114

## Geometric Algebra equivalants for Pauli Matrices.

Posted by peeterjoot on July 12, 2009

# Motivation.

Having learned Geometric (Clifford) Algebra from ([1]), ([2]), ([3]), and other sources before studying any quantum mechanics, trying to work with (and talk to people familiar with) the Pauli and Dirac matrix notation as used in traditional quantum mechanics becomes difficult.

The aim of these notes is to work through equivalents to many Clifford algebra expressions entirely in commutator and anticommutator notations. This will show the mapping between the (generalized) dot product and the wedge product, and also show how the different grade elements of the Clifford algebra $\mathcal{C}\mathcal{L}(3,0)$ manifest in their matrix forms.

# Pauli Matrices

The matrices in question are:

\begin{aligned}\sigma_1 &= \begin{bmatrix} 0 & 1 \\ 1 & 0 \\ \end{bmatrix} \\ \sigma_2 &= \begin{bmatrix} 0 & -i \\ i & 0 \\ \end{bmatrix} \\ \sigma_3 &= \begin{bmatrix} 1 & 0 \\ 0 & -1 \\ \end{bmatrix} \end{aligned}

These all have positive square as do the traditional Euclidean unit vectors $\mathbf{e}_i$, and so can be used algebraically as a vector basis for $\mathbb{R}^{3}$. So any vector that we can write in coordinates

\begin{aligned}\mathbf{x} = x^i \mathbf{e}_i, \end{aligned}

we can equivalently write (an isomorphism) in terms of the Pauli matrix’s

\begin{aligned}x = x^i \sigma_i. \end{aligned} \quad\quad\quad(4)

## Pauli Vector.

([4]) introduces the Pauli vector as a mechanism for mapping between a vector basis and this matrix basis

\begin{aligned}\boldsymbol{\sigma} = \sum \sigma_i \mathbf{e}_i \end{aligned}

This is a curious looking construct with products of $2 x 2$ matrices and $\mathbb{R}^{3}$ vectors. Obviously these are not the usual $3 x 1$ column vector representations. This Pauli vector is thus really a notational construct. If one takes the dot product of a vector expressed using the standard orthonormal Euclidean basis $\{\mathbf{e}_i\}$ basis, and then takes the dot product with the Pauli matrix in a mechanical fashion

\begin{aligned}\mathbf{x} \cdot \boldsymbol{\sigma} &=(x^i \mathbf{e}_i) \cdot \sum \sigma_j \mathbf{e}_j \\ &= \sum_{i,j} x^i \sigma_j \mathbf{e}_i \cdot \mathbf{e}_j \\ &= x^i \sigma_i \end{aligned}

one arrives at the matrix representation of the vector in the Pauli basis $\{\sigma_i\}$. Does this construct have any value? That I don’t know, but for the rest of these notes the coordinate representation as in equation (4) will be used directly.

## Matrix squares

It was stated that the Pauli matrices have unit square. Direct calculation of this is straightforward, and confirms the assertion

\begin{aligned}{\sigma_1}^2 &= \begin{bmatrix} 0 & 1 \\ 1 & 0 \\ \end{bmatrix} \begin{bmatrix} 0 & 1 \\ 1 & 0 \\ \end{bmatrix} = \begin{bmatrix} 1 & 0 \\ 0 & 1 \\ \end{bmatrix} = I \\ {\sigma_2}^2 &= \begin{bmatrix} 0 & -i \\ i & 0 \\ \end{bmatrix} \begin{bmatrix} 0 & -i \\ i & 0 \\ \end{bmatrix} = i^2 \begin{bmatrix} 0 & -1 \\ 1 & 0 \\ \end{bmatrix} \begin{bmatrix} 0 & -1 \\ 1 & 0 \\ \end{bmatrix} = \begin{bmatrix} 1 & 0 \\ 0 & 1 \\ \end{bmatrix} = I \\ {\sigma_3}^2 &= \begin{bmatrix} 1 & 0 \\ 0 & -1 \\ \end{bmatrix} \begin{bmatrix} 1 & 0 \\ 0 & -1 \\ \end{bmatrix} = \begin{bmatrix} 1 & 0 \\ 0 & 1 \\ \end{bmatrix} = I \end{aligned}

Note that unlike the vector (Clifford) square the identity matrix and not a scalar.

## Length

If we are to operate with Pauli matrices how do we express our most basic vector operation, the length?

Examining a vector lying along one direction, say, $\mathbf{a} = \alpha\hat{\mathbf{x}}$ we expect

\begin{aligned}\mathbf{a}^2 = \mathbf{a} \cdot \mathbf{a} = \alpha^2 \hat{\mathbf{x}} \cdot \hat{\mathbf{x}} = \alpha^2. \end{aligned}

Lets contrast this to the Pauli square for the same vector $y = \alpha\sigma_1$

\begin{aligned}y^2 = \alpha^2 {\sigma_1}^2 = \alpha^2 I \end{aligned}

The wiki article mentions trace, but no application for it. Since $\text{Tr}\left({{I}}\right) = 2$, an observable application is that the trace operator provides a mechanism to convert a diagonal matrix to a scalar. In particular for this scaled unit vector $y$ we have

\begin{aligned}\alpha^2 = \frac{1}{{2}} \text{Tr}\left({{y^2}}\right) \end{aligned}

It is plausible to guess that the squared length will be related to the matrix square in the general case as well

\begin{aligned}{\left\lvert{x}\right\rvert}^2 = \frac{1}{{2}}\text{Tr}\left({{x^2}}\right) \end{aligned}

Let’s see if this works by performing the coordinate expansion

\begin{aligned}x^2&= (x^i \sigma_i)(x^j \sigma_j) \\ &= x^i x^j \sigma_i \sigma_j \end{aligned}

A split into equal and different indexes thus leaves

\begin{aligned}x^2&= \sum_{i < j} x^i x^j (\sigma_i \sigma_j + \sigma_j \sigma_i) + \sum_i (x^i)^2 {\sigma_i}^2 \end{aligned} \quad\quad\quad(5)

As an algebra that is isomorphic to the Clifford Algebra $\mathcal{C}\mathcal{L}_{\{{3},{0}\}}$ it is expected that the $\sigma_i \sigma_j$ matrices anticommute for $i \ne j$. Multiplying these out verifies this

\begin{aligned}\begin{array}{l l l l}\sigma_1 \sigma_2 &= i \begin{bmatrix} 0 & 1 \\ 1 & 0 \\ \end{bmatrix} \begin{bmatrix} 0 & -1 \\ 1 & 0 \\ \end{bmatrix} &= i \begin{bmatrix} 1 & 0 \\ 0 & -1 \\ \end{bmatrix} &= i \sigma_3 \\ \sigma_2 \sigma_1 &= i \begin{bmatrix} 0 & -1 \\ 1 & 0 \\ \end{bmatrix} \begin{bmatrix} 0 & 1 \\ 1 & 0 \\ \end{bmatrix} &= i \begin{bmatrix}-1 & 0 \\ 0 & 1\end{bmatrix} &= -i \sigma_3 \\ \sigma_3 \sigma_1 &= \begin{bmatrix} 1 & 0 \\ 0 & -1 \\ \end{bmatrix} \begin{bmatrix} 0 & 1 \\ 1 & 0 \\ \end{bmatrix} &= \begin{bmatrix}0 & 1 \\ -1 & 0\end{bmatrix} &= i \sigma_2 \\ \sigma_1 \sigma_3 &= \begin{bmatrix} 0 & 1 \\ 1 & 0 \\ \end{bmatrix} \begin{bmatrix} 1 & 0 \\ 0 & -1 \\ \end{bmatrix} &= \begin{bmatrix}0 & -1 \\ 1 & 0\end{bmatrix} &= -i \sigma_2 \\ \sigma_2 \sigma_3 &= i \begin{bmatrix} 0 & -1 \\ 1 & 0 \\ \end{bmatrix} \begin{bmatrix} 1 & 0 \\ 0 & -1 \\ \end{bmatrix} &= i \begin{bmatrix} 0 & 1 \\ 1 & 0 \\ \end{bmatrix} &= i \sigma_1 \\ \sigma_3 \sigma_2 &= i \begin{bmatrix} 1 & 0 \\ 0 & -1 \\ \end{bmatrix} \begin{bmatrix} 0 & -1 \\ 1 & 0 \\ \end{bmatrix} &= i \begin{bmatrix}0 & -1 \\ -1 & 0\end{bmatrix} &= -i \sigma_3 \\ \end{array}. \end{aligned}

Thus in (5) the sum over the $\{i < j\} = \{12, 23, 13\}$ indexes is zero.

Having computed this, our vector square leaves us with the vector length multiplied by the identity matrix

\begin{aligned}x^2 = \sum_i (x^i)^2 I. \end{aligned}

Invoking the trace operator will therefore extract just the scalar length desired

\begin{aligned}{\left\lvert{x}\right\rvert}^2 = \frac{1}{{2}} \text{Tr}\left({{x^2}}\right) = \sum_i (x^i)^2. \end{aligned}

### Aside: Summarizing the multiplication table.

It is worth pointing out that the multiplication table above used to confirm the antisymmetric behavior of the Pauli basis can be summarized as

\begin{aligned}\sigma_a \sigma_b = 2 i \epsilon_{abc} \sigma_c \end{aligned}

## Scalar product.

Having found the expression for the length of a vector in the Pauli basis, the next logical desirable identity is the dot product. One can guess that this will be the trace of a scaled symmetric product, but can motivate this without guessing in the usual fashion, by calculating the length of an orthonormal sum.

Consider first the length of a general vector sum. To calculate this we first wish to calculate the matrix square of this sum.

\begin{aligned}(x + y)^2 &= x^2 + y^2 + x y + y x \end{aligned}

If these vectors are perpendicular this equals $x^2 + y^2$. Thus orthonormality implies that

\begin{aligned}x y + y x &= 0 \end{aligned}

or,

\begin{aligned}y x &= - x y \end{aligned}

We have already observed this by direct calculation for the Pauli matrices themselves. Now, this is not any different than the usual description of perpendicularity in a Clifford Algebra, and it is notable that there are not any references to matrices in this argument. One only requires that a well defined vector product exists, where the squared vector has a length interpretation.

One matrix dependent observation that can be made is that since the left hand side and the $x^2$, and $y^2$ terms are all diagonal, this symmetric sum must also be diagonal. Additionally, for the length of this vector sum we then have

\begin{aligned}{{\left\lvert{x + y}\right\rvert}}^2&= {\left\lvert{x}\right\rvert}^2 + {\left\lvert{y}\right\rvert}^2 + \frac{1}{{2}}\text{Tr}\left({{x y + y x}}\right) \end{aligned}

For correspondence with the Euclidean dot product of two vectors we must then have

\begin{aligned}{x} \bullet {y} &= \frac{1}{{4}}\text{Tr}\left({{x y + y x}}\right). \end{aligned}

Here $x \bullet y$ has been used to denote this scalar product (ie: a plain old number), since $x \cdot y$ will be used later for a matrix dot product (this times the identity matrix) which is more natural in many ways for this Pauli algebra.

Observe the symmetric product that is found embedded in this scalar selection operation. In physics this is known as the anticommutator, where the commutator is the antisymmetric sum. In the physics notation the anticommutator (symmetric sum) is

\begin{aligned}\left\{{x},{y}\right\} &= x y + y x \end{aligned} \quad\quad\quad(10)

So this scalar selection can be written

\begin{aligned}{ x} \bullet { y} = \frac{1}{{4}}\text{Tr}{{\left\{{x},{y}\right\}}} \end{aligned}

Similarly, the commutator, an antisymmetric product, is denoted:

\begin{aligned}\left[{x},{y}\right] &= x y - y x, \end{aligned} \quad\quad\quad(12)

A close relationship between this commutator and the wedge product of Clifford Algebra is expected.

## Symmetric and antisymmetric split.

As with the Clifford product, the symmetric and antisymmetric split of a vector product is a useful concept. This can be used to write the product of two Pauli basis vectors in terms of the anticommutator and commutator products

\begin{aligned}x y &= \frac{1}{{2}} \left\{{x},{y}\right\} + \frac{1}{{2}} \left[{x},{y}\right] \\ y x &= \frac{1}{{2}} \left\{{x},{y}\right\} - \frac{1}{{2}} \left[{x},{y}\right] \end{aligned}

These follows from the definition of the anticommutator (10) and commutator (12) products above, and are the equivalents of the Clifford symmetric and antisymmetric split into dot and wedge products

\begin{aligned}x y &= {x} \cdot {y} + {x} \wedge {y} \\ y x &= {x} \cdot {y} - {x} \wedge {y} \end{aligned}

Where the dot and wedge products are respectively

\begin{aligned}x \cdot y &= \frac{1}{{2}}(x y + y x) \\ x \wedge y &= \frac{1}{{2}}(x y - y x) \end{aligned}

Note the factor of two differences in the two algebraic notations. In particular very handy Clifford vector product reversal formula

\begin{aligned}y x = - x y + 2 x \cdot y \end{aligned}

has no factor of two in its Pauli anticommutator equivalent

\begin{aligned}y x = - x y + \left\{{x},{y}\right\} \end{aligned}

## Vector inverse.

It has been observed that the square of a vector is diagonal in this matrix representation, and can therefore be inverted for any non-zero vector

\begin{aligned}x^2 &= {\left\lvert{x}\right\rvert}^2 I \\ (x^2)^{-1} &= {\left\lvert{x}\right\rvert}^{-2} I \\ \implies \\ x^2 (x^2)^{-1} &= I \end{aligned}

So it is therefore quite justifiable to define

\begin{aligned}x^{-2} = \frac{1}{{x^2}} \equiv {\left\lvert{x}\right\rvert}^{-2} I \end{aligned}

This allows for the construction of a dual sided vector inverse operation.

\begin{aligned}x^{-1}&\equiv \frac{1}{{{\left\lvert{x}\right\rvert}^2}} x \\ &= \frac{1}{{x^2}} x \\ &= x \frac{1}{{x^2}} \end{aligned}

This inverse is a scaled version of the vector itself.

The diagonality of the squared matrix or the inverse of that allows for commutation with x. This diagonality plays the same role as the scalar in a regular Clifford square. In either case the square can commute with the vector, and that commutation allows the inverse to have both left and right sided action.

Note that like the Clifford vector inverse when the vector is multiplied with this inverse, the product resides outside of the proper $\mathbb{R}^{3}$ Pauli basis since the identity matrix is required.

## Coordinate extraction.

Given a vector in the Pauli basis, we can extract the coordinates using the scalar product

\begin{aligned}x = \sum_i \frac{1}{{4}}\text{Tr}{{\left\{{x},{\sigma_i}\right\}}} \sigma_i \end{aligned}

But do not need to convert to strict scalar form if we are multiplying by a Pauli matrix. So in anticommutator notation this takes the form

\begin{aligned}x &= x^i \sigma_i = \sum_i \frac{1}{{2}}\left\{{x},{\sigma_i}\right\} \sigma_i \\ x^i &= \frac{1}{{2}}\left\{{x},{\sigma_i}\right\} \end{aligned} \quad\quad\quad(18)

## Projection and rejection.

The usual Clifford algebra trick for projective and rejective split maps naturally to matrix form. Write

\begin{aligned}x &= x a a^{-1} \\ &= (x a) a^{-1} \\ &= \left( \frac{1}{{2}}\left\{{x},{a}\right\} + \frac{1}{{2}}\left[{x},{a}\right] \right) a^{-1} \\ &= \left( \frac{1}{{2}}\left(x a + a x \right) + \frac{1}{{2}} \left(x a - a x \right) \right) a^{-1} \\ &= \frac{1}{{2}}\left(x + a x a^{-1} \right) + \frac{1}{{2}} \left(x - a x a^{-1} \right) \end{aligned}

Since $\left\{{x},{a}\right\}$ is diagonal, this first term is proportional to $a^{-1}$, and thus lines in the direction of $a$ itself. The second term is perpendicular to $a$.

These are in fact the projection of $x$ in the direction of $a$ and rejection of $x$ from the direction of $a$ respectively.

\begin{aligned}x &= x_\parallel + x_\perp \\ x_\parallel &= \text{Proj}_a(x) = \frac{1}{{2}}\left\{{x},{a}\right\}a^{-1} = \frac{1}{{2}}\left(x + a x a^{-1} \right) \\ x_\perp &= \text{Rej}_a(x) = \frac{1}{{2}} \left[{x},{a}\right] a^{-1} = \frac{1}{{2}} \left( x - a x a^{-1} \right) \end{aligned}

To complete the verification of this note that the perpendicularity of the $x_\perp$ term can be verified by taking dot products

\begin{aligned}\frac{1}{{2}}\left\{{a},{x_\perp}\right\} &= \frac{1}{{4}}\left( a \left(x - a x a^{-1} \right) +\left(x - a x a^{-1} \right) a \right) \\ &= \frac{1}{{4}}\left( a x - a a x a^{-1} + x a - a x a^{-1} a \right) \\ &= \frac{1}{{4}}\left( a x - x a + x a - a x \right) \\ &= 0 \end{aligned}

## Space of the vector product.

Expansion of the anticommutator and commutator in coordinate form shows that these entities lie in a different space than the vectors itself.

For real coordinate vectors in the Pauli basis, all the commutator values are imaginary multiples and thus not representable

\begin{aligned}\left[{x},{y}\right] &= x^a \sigma_a y^b \sigma_b - y^a \sigma_a x^b \sigma_b \\ &= (x^a y^b - y^a x^b) \sigma_a \sigma_b \\ &= 2 i (x^a y^b - y^a x^b) \epsilon_{a b c} \sigma_c \end{aligned}

Similarly, the anticommutator is diagonal, which also falls outside the Pauli vector basis:

\begin{aligned}\left\{{x},{y}\right\} &= x^a \sigma_a y^b \sigma_b + y^a \sigma_a x^b \sigma_b \\ &= (x^a y^b + y^a x^b) \sigma_a \sigma_b \\ &= (x^a y^b + y^a x^b) ( I \delta_{a b} + i \epsilon_{a b c} \sigma_c) \\ &= \sum_a (x^a y^a + y^a x^a) I +\sum_{a

These correspond to the Clifford dot product being scalar (grade zero), and the wedge defining a grade two space, where grade expresses the minimal degree that a product can be reduced to. By example a Clifford product of normal unit vectors such as

\begin{aligned}\mathbf{e}_1 \mathbf{e}_3 \mathbf{e}_4 \mathbf{e}_1 \mathbf{e}_3 \mathbf{e}_4 \mathbf{e}_3 &\propto \mathbf{e}_3 \\ \mathbf{e}_2 \mathbf{e}_3 \mathbf{e}_4 \mathbf{e}_1 \mathbf{e}_3 \mathbf{e}_4 \mathbf{e}_3 \mathbf{e}_5 &\propto \mathbf{e}_1 \mathbf{e}_2 \mathbf{e}_3 \mathbf{e}_5 \end{aligned}

are grade one and four respectively. The proportionality constant will be dependent on metric of the underlying vector space and the number of permutations required to group terms in pairs of matching indexes.

## Completely antisymmetrized product of three vectors.

In a Clifford algebra no imaginary number is required to express the antisymmetric (commutator) product. However, the bivector space can be enumerated using a dual basis defined by multiplication of the vector basis elements with the unit volume trivector. That is also the case here and gives a geometrical meaning to the imaginaries of the Pauli formulation.

How do we even write the unit volume element in Pauli notation? This would be

\begin{aligned}\sigma_1 \wedge \sigma_2 \wedge \sigma_3 &= (\sigma_1 \wedge \sigma_2) \wedge \sigma_3 \\ &= \frac{1}{{2}} \left[{\sigma_1},{\sigma_2}\right] \wedge \sigma_3 \\ &= \frac{1}{{4}} \left( \left[{\sigma_1},{\sigma_2}\right] \sigma_3 + \sigma_3 \left[{\sigma_1},{\sigma_2}\right] \right) \end{aligned}

So we have

\begin{aligned}\sigma_1 \wedge \sigma_2 \wedge \sigma_3 &= \frac{1}{{8}} \left\{\left[{\sigma_1},{\sigma_2}\right],{\sigma_3}\right\} \end{aligned} \quad\quad\quad(20)

Similar expansion of $\sigma_1 \wedge \sigma_2 \wedge \sigma_3 = \sigma_1 \wedge (\sigma_2 \wedge \sigma_3)$, or $\sigma_1 \wedge \sigma_2 \wedge \sigma_3 = (\sigma_3 \wedge \sigma_1) \wedge \sigma_2$ shows that we must also have

\begin{aligned}\left\{\left[{\sigma_1},{\sigma_2}\right],{\sigma_3}\right\}= \left\{{\sigma_1},\left[{\sigma_2},{\sigma_3}\right]\right\}= \left\{\left[{\sigma_3},{\sigma_1}\right],{\sigma_2}\right\} \end{aligned}

Until now the differences in notation between the anticommutator/commutator and the dot/wedge product of the Pauli algebra and Clifford algebra respectively have only differed by factors of two, which isn’t much of a big deal. However, having to express the naturally associative wedge product operation in the non-associative looking notation of equation (20) is rather unpleasant seeming. Looking at an expression of the form gives no mnemonic hint of the underlying associativity, and actually seems obfuscating. I suppose that one could get used to it though.

We expect to get a three by three determinant out of the trivector product. Let’s verify this by expanding this in Pauli notation for three general coordinate vectors

\begin{aligned}\left\{\left[{x},{y}\right],{z}\right\} &= \left\{\left[{x^a \sigma_a},{y^b \sigma_b}\right],{ z^c \sigma_c}\right\} \\ &= 2 i \epsilon_{a b d} x^a y^b z^c \left\{{\sigma_d},{\sigma_c}\right\} \\ &= 4 i \epsilon_{a b d} x^a y^b z^c \delta_{c d} I \\ &= 4 i \epsilon_{a b c} x^a y^b z^c I \\ &= 4 i \begin{vmatrix} x_a & x_b & x_c \\ y_a & y_b & y_c \\ z_a & z_b & z_c\end{vmatrix} I \end{aligned}

In particular, our unit volume element is

\begin{aligned}\sigma_1 \wedge \sigma_2 \wedge \sigma_3 &= \frac{1}{{4}} \left\{\left[{\sigma_1},{\sigma_2}\right],{\sigma_3}\right\} = i I \end{aligned}

So one sees that the complex number $i$ in the Pauli algebra can logically be replaced by the unit pseudoscalar $i I$, and relations involving $i$, like the commutator expansion of a vector product, is restored to the expected dual form of Clifford algebra

\begin{aligned}\sigma_a \wedge \sigma_b &= \frac{1}{{2}} \left[{\sigma_a},{\sigma_b}\right] \\ &= i \epsilon_{a b c} \sigma_c \\ &= (\sigma_a \wedge \sigma_b \wedge \sigma_c) \sigma_c \end{aligned}

Or

\begin{aligned}\sigma_a \wedge \sigma_b &= (\sigma_a \wedge \sigma_b \wedge \sigma_c) \cdot \sigma_c \end{aligned} \quad\quad\quad(23)

## Duality.

We’ve seen that multiplication by $i$ is a duality operation, which is expected since $iI$ is the matrix equivalent of the unit pseudoscalar. Logically this means that for a vector $x$, the product $(iI) x$ represents a plane quantity (torque, angular velocity/momentum, …). Similarly if $B$ is a plane object, then $(iI) B$ will have a vector interpretation.

In particular, for the antisymmetric (commutator) part of the vector product $x y$

\begin{aligned}\frac{1}{{2}} \left[{x},{y}\right]&= \frac{1}{{2}} x^a y^b \left[{\sigma_a},{\sigma_b}\right] \\ &= x^a y^b i \epsilon_{a b c}{\sigma_c} \end{aligned}

a “vector” in the dual space spanned by $\{i \sigma_a\}$ is seen to be more naturally interpreted as a plane quantity (bivector in Clifford algebra).

As in Clifford algebra, we can write the cross product in terms of the antisymmetric product

\begin{aligned}a \times b = \frac{1}{{2i}} \left[{a},{b}\right]. \end{aligned}

With the factor of $2$ in the denominator here (like the exponential form of sine), it is interesting
to contrast this to the cross product in its trigonometric form

\begin{aligned}a \times b &= {\left\lvert{a}\right\rvert}{\left\lvert{b}\right\rvert} \sin(\theta) \hat{\mathbf{n}} \\ &= {\left\lvert{a}\right\rvert}{\left\lvert{b}\right\rvert} \frac{1}{{2i}} ( e^{i\theta} - e^{-i\theta}) \hat{\mathbf{n}} \end{aligned}

This shows we can make the curious identity

\begin{aligned}\left[{\hat{a}},{\hat{b}}\right] &= ( e^{i\theta} - e^{-i\theta}) \hat{\mathbf{n}} \end{aligned}

If one did not already know about the dual sides half angle rotation formulation of Clifford algebra, this is a hint about how one could potentially work towards that. We have the commutator (or wedge product) as a rotation operator that leaves the normal component of a vector untouched (commutes with the normal vector).

## Complete algebraic space.

Pauli equivalents for all the elements in the Clifford algebra have now been
determined.

\begin{itemize}
\item scalar

\begin{aligned}\alpha \rightarrow \alpha I \end{aligned}

\item vector

\begin{aligned}u^i \sigma_i &\rightarrow \begin{bmatrix}0 & u^1 \\ u^1 & 0 \\ \end{bmatrix}+\begin{bmatrix}0 & -i u^2 \\ i u^2 & 0 \\ \end{bmatrix}+\begin{bmatrix}u^3 & 0 \\ 0 & -u^3 \\ \end{bmatrix} \\ &=\begin{bmatrix}u^3 & u^1 -i u^2 \\ u^1 + i u^2 & -u^3 \\ \end{bmatrix} \end{aligned}

\item bivector

\begin{aligned}\sigma_1\sigma_2\sigma_3 v^a \sigma_a &\rightarrow i v^a \sigma_a \\ &= \begin{bmatrix}iv^3 & iv^1 + v^2 \\ iv^1 - v^2 & -i v^3 \\ \end{bmatrix} \end{aligned}

\item pseudoscalar

\begin{aligned}\beta \sigma_1\sigma_2\sigma_3 \rightarrow i \beta I \end{aligned}

\end{itemize}

Summing these we have the mapping from Clifford basis to Pauli matrix as follows

\begin{aligned}\alpha + \beta I+ u^i \sigma_i+ I v^a \sigma_a &\rightarrow \begin{bmatrix}(\alpha + u^3) + i(\beta + v^3) & (u^1 + v^2) +i (-u^2 + v^1) \\ (u^1 - v^2) + i (u^2 - v^1) & (\alpha -u^3) + i(\beta - v^3)\\ \end{bmatrix} \end{aligned}

Thus for any given sum of scalar, vector, bivector, and trivector elements we can completely express this in Pauli form as a general $2 x 2$ complex matrix.

Provided that one can also extract the coordinates for each of the grades involved, this also provides a complete Clifford algebra characterization of an arbitrary complex $2 x 2$ matrix.

Computationally this has some nice looking advantages. Given any canned complex matrix software, one should be able to easily cook up with little work a working $\mathbb{R}^{3}$ Clifford calculator.

As for the coordinate extraction, part of the work can be done by taking real and imaginary components. Let an element of the general algebra be denoted

\begin{aligned}P = \begin{bmatrix}z_{11} & z_{12} \\ z_{21} & z_{22} \\ \end{bmatrix} \end{aligned}

We therefore have

\begin{aligned}\Re(P) &=\begin{bmatrix}\alpha + u^3 & u^1 + v^2 \\ u^1 - v^2 & \alpha -u^3 \\ \end{bmatrix} \\ \Im(P) &=\begin{bmatrix}\beta + v^3 & -u^2 + v^1 \\ u^2 + v^1 & \beta - v^3 \\ \end{bmatrix} \end{aligned}

By inspection, symmetric and antisymmetric sums of the real and imaginary parts recovers the coordinates as follows

\begin{aligned}\alpha &= \frac{1}{{2}} \Re( z_{11} + z_{22} ) \\ u^3 &= \frac{1}{{2}} \Re( z_{11} - z_{22} ) \\ u^1 &= \frac{1}{{2}} \Re( z_{12} + z_{21} ) \\ v^2 &= \frac{1}{{2}} \Re( z_{12} - z_{21} ) \\ \beta &= \frac{1}{{2}} \Im( z_{11} + z_{22} ) \\ v^3 &= \frac{1}{{2}} \Im( z_{11} - z_{22} ) \\ v^1 &= \frac{1}{{2}} \Im( z_{12} + z_{21} ) \\ u^2 &= \frac{1}{{2}} \Im( -z_{12} + z_{21} ) \end{aligned}

\begin{aligned}P = \left\langle{{P}}\right\rangle +{\left\langle{{P}}\right\rangle}_{1} +{\left\langle{{P}}\right\rangle}_{2} +{\left\langle{{P}}\right\rangle}_{3}, \end{aligned}

is

\begin{aligned}\left\langle{{P}}\right\rangle &= \frac{1}{{2}} \Re( z_{11} + z_{22} ) = \frac{1}{{2}} \Re(\text{Tr}ace P) \\ {\left\langle{{P}}\right\rangle}_{1} &= \frac{1}{{2}} \left(\Re( z_{12} + z_{21} ) \sigma_1+\Im( -z_{12} + z_{21} ) \sigma_2+\Re( z_{11} - z_{22} ) \sigma_3\right) \\ {\left\langle{{P}}\right\rangle}_{2}&= \frac{1}{{2}} \left(\Im( z_{12} + z_{21} ) \sigma_2 \wedge \sigma_3+\Re( z_{12} - z_{21} ) \sigma_3 \wedge \sigma_1+\Im( z_{11} - z_{22} ) \sigma_1 \wedge \sigma_2\right) \\ {\left\langle{{P}}\right\rangle}_{3} &= \frac{1}{{2}} \Im( z_{11} + z_{22} ) I = \frac{1}{{2}} \Im(\text{Tr}ace P) \sigma_1 \wedge \sigma_2 \wedge \sigma_3 \end{aligned}

Employing $\Im(z) = \Re(-iz)$, and $\Re(z) = \Im(iz)$ this can be made slightly more symmetrical, with Real operations selecting the vector coordinates and imaginary operations selecting the bivector coordinates.

\begin{aligned}\left\langle{{P}}\right\rangle &= \frac{1}{{2}} \Re( z_{11} + z_{22} ) = \frac{1}{{2}} \Re(\text{Tr}ace P) \\ {\left\langle{{P}}\right\rangle}_{1} &= \frac{1}{{2}} \left(\Re( z_{12} + z_{21} ) \sigma_1+\Re( iz_{12} - iz_{21} ) \sigma_2+\Re( z_{11} - z_{22} ) \sigma_3\right) \\ {\left\langle{{P}}\right\rangle}_{2}&= \frac{1}{{2}} \left(\Im( z_{12} + z_{21} ) \sigma_2 \wedge \sigma_3+\Im( iz_{12} - iz_{21} ) \sigma_3 \wedge \sigma_1+\Im( z_{11} - z_{22} ) \sigma_1 \wedge \sigma_2\right) \\ {\left\langle{{P}}\right\rangle}_{3} &= \frac{1}{{2}} \Im( z_{11} + z_{22} ) I = \frac{1}{{2}} \Im(\text{Tr}ace P) \sigma_1 \wedge \sigma_2 \wedge \sigma_3 \end{aligned}

Finally, returning to the Pauli algebra, this also provides the following split of the Pauli multivector matrix into its geometrically significant components $P = \left\langle{{P}}\right\rangle +{\left\langle{{P}}\right\rangle}_{1} +{\left\langle{{P}}\right\rangle}_{2} +{\left\langle{{P}}\right\rangle}_{3}$,

\begin{aligned}\left\langle{{P}}\right\rangle &= \frac{1}{{2}} \Re( z_{11} + z_{22} ) I \\ {\left\langle{{P}}\right\rangle}_{1} &= \frac{1}{{2}} \left(\Re( z_{12} + z_{21} ) \sigma_1+\Re( iz_{12} - iz_{21} ) \sigma_2+\Re( z_{11} - z_{22} ) \sigma_3\right) \\ {\left\langle{{P}}\right\rangle}_{2}&= \frac{1}{{2}} \left(\Im( z_{12} + z_{21} ) i\sigma_1+\Im( iz_{12} - iz_{21} ) i\sigma_2+\Im( z_{11} - z_{22} ) i\sigma_k\right) \\ {\left\langle{{P}}\right\rangle}_{3} &= \frac{1}{{2}} \Im( z_{11} + z_{22} ) iI \end{aligned}

## Reverse operation.

The reversal operation switches the order of the product of perpendicular vectors. This will change the sign of grade two and three terms in the Pauli algebra. Since $\sigma_2$ is imaginary, conjugation does not have the desired effect, but Hermitian conjugation (conjugate transpose) does the trick.

Since the reverse operation can be written as Hermitian conjugation, one can also define the anticommutator and commutator in terms of reversion in a way that seems particularly natural for complex matrices. That is

\begin{aligned}\left\{{a},{b}\right\} &= ab + (ab)^{*} \\ \left[{a},{b}\right] &= ab - (ab)^{*} \end{aligned}

## Rotations.

Rotations take the normal Clifford, dual sided quaterionic form. A rotation about a unit normal $n$ will be

\begin{aligned}R(x) = e^{-i n \theta/2} x e^{i n \theta/2} \end{aligned}

The Rotor $R = e^{-i n \theta/2}$ commutes with any component of the vector x that is parallel to the normal (perpendicular to the plane), whereas it anticommutes with the components
in the plane. Writing the vector components perpendicular and parallel to the plane respectively as $x = x_\perp + x_\parallel$, the essence of the rotation action is this selective commutation or anti-commutation behavior

\begin{aligned}R x_\parallel R^{*} &= x_\parallel R^{*} \\ R x_\perp R^{*} &= x_\perp R R^{*} = x_\perp \end{aligned}

Here the exponential has the obvious meaning in terms of exponential series, so for this bivector case we have

\begin{aligned}\exp(i\hat{\mathbf{n}}\theta/2) &= \cos(\theta/2) I + i\hat{\mathbf{n}} \sin(\theta/2) \end{aligned}

The unit bivector $B = i\hat{\mathbf{n}}$ can also be defined explicitly in terms of two vectors $a$, and $b$ in the plane

\begin{aligned}B = \frac{1}{{{\left\lvert{\left[{a},{b}\right]}\right\rvert}}} \left[{a},{b}\right] \end{aligned}

Where the bivector length is defined in terms of the conjugate square (bivector times bivector reverse)

\begin{aligned}{\left\lvert{\left[{a},{b}\right]}\right\rvert}^2 = \left[{a},{b}\right] {\left[{a},{b}\right]}^{*} \end{aligned}

Examples to complete this subsection would make sense. As one of the most powerful and useful operations in the algebra, it is a shame in terms of completeness to skimp on this. However, except for some minor differences like substitution of the Hermitian conjugate operation for reversal, the use of the identity matrix $I$ in place of the scalar in the exponential expansion, the treatment is exactly the same as in the Clifford algebra.

Coordinate equations for grade selection were worked out above, but the observation that reversion and Hermitian conjugation are isomorphic operations can partially clean this up. In particular a Hermitian conjugate symmetrization and anti-symmetrization of the general matrix provides a nice split into quaternion and dual quaternion parts (say $P = Q + R$ respectively). That is

\begin{aligned}Q &= \left\langle{{P}}\right\rangle + {\left\langle{{P}}\right\rangle}_{1} = \frac{1}{{2}}(P + P^{*}) \\ R &= {\left\langle{{P}}\right\rangle}_{2} + {\left\langle{{P}}\right\rangle}_{3} = \frac{1}{{2}}(P - P^{*}) \end{aligned}

Now, having done that, how to determine $\left\langle{{Q}}\right\rangle$, ${\left\langle{{Q}}\right\rangle}_{1}$, ${\left\langle{{R}}\right\rangle}_{2}$, and ${\left\langle{{R}}\right\rangle}_{3}$ becomes the next question. Once that is done, the individual coordinates can be picked off easily enough. For the vector parts, a Fourier decomposition as in equation (18) will retrieve the desired coordinates.

The dual vector coordinates can be picked off easily enough taking dot products with the dual basis vectors

\begin{aligned}B &= B^k i \sigma_k = \sum_k \frac{1}{{2}}\left\{{B},{\frac{1}{{i \sigma_k}}}\right\} i\sigma_k \\ B^k &= \frac{1}{{2}}\left\{{B},{\frac{1}{{i\sigma_k}}}\right\} \end{aligned} \quad\quad\quad(27)

For the quaternion part $Q$ the aim is to figure out how to isolate or subtract out the scalar part. This is the only tricky bit because the diagonal bits are all mixed up with the $\sigma_3$ term which is also real, and diagonal. Consideration of the sum

\begin{aligned}a I + b \sigma_3 = \begin{bmatrix}a + b & 0 \\ 0 & a - b \\ \end{bmatrix}, \end{aligned}

shows that trace will recover the value $2a$, so we have

\begin{aligned}\left\langle{{Q}}\right\rangle &= \frac{1}{{2}}\text{Tr}\left({{Q}}\right) I \\ {\left\langle{{Q}}\right\rangle}_{1} &= Q - \frac{1}{{2}}\text{Tr}\left({{Q}}\right) I. \end{aligned}

Next is isolation of the pseudoscalar part of the dual quaternion $R$. As with the scalar part, consideration of the sum of the $i\sigma_3$ term and the $iI$ term is required

\begin{aligned}ia I + ib \sigma_3 = \begin{bmatrix}ia + ib & 0 \\ 0 & ia - ib \\ \end{bmatrix}, \end{aligned}

So the trace of the dual quaternion provides the $2a$, leaving the bivector and pseudoscalar grade split

\begin{aligned}{\left\langle{{R}}\right\rangle}_{3} &= \frac{1}{{2}}\text{Tr}\left({{R}}\right) I \\ {\left\langle{{R}}\right\rangle}_{2} &= R - \frac{1}{{2}}\text{Tr}\left({{R}}\right) I. \end{aligned}

A final assembly of these results provides the following coordinate free grade selection operators

\begin{aligned}\left\langle{{P}}\right\rangle &= \frac{1}{{4}}\text{Tr}\left({{P + P^{*}}}\right) I \\ {\left\langle{{P}}\right\rangle}_{1} &= \frac{1}{{2}} (P + P^{*}) - \frac{1}{{4}}\text{Tr}\left({{P + P^{*}}}\right) I \\ {\left\langle{{P}}\right\rangle}_{2} &= \frac{1}{{2}} (P - P^{*}) - \frac{1}{{4}}\text{Tr}\left({{P - P^{*}}}\right) I \\ {\left\langle{{P}}\right\rangle}_{3} &= \frac{1}{{4}}\text{Tr}\left({{P - P^{*}}}\right) I \end{aligned}

## Generalized dot products.

Here the equivalent of the generalized Clifford bivector/vector dot product will be computed, as well as the associated distribution equation

\begin{aligned}( a \wedge b ) \cdot c &= a (b \cdot c) - b (a \cdot c) \end{aligned}

To translate this write

\begin{aligned}( a \wedge b ) \cdot c &= \frac{1}{{2}} \left( (a \wedge b) c - c (a \wedge b) \right) \end{aligned} \quad\quad\quad(33)

Then with the identifications

\begin{aligned}a \cdot b &\equiv \frac{1}{{2}} \left\{{a},{b}\right\} \\ a \wedge b &\equiv \frac{1}{{2}} \left[{a},{b}\right] \end{aligned}

we have

\begin{aligned}( a \wedge b ) \cdot c &\equiv \frac{1}{{4}} \left[{\left[{a},{b}\right]},{c}\right] \\ &= \frac{1}{{2}} \left( a \left\{{b},{c}\right\} - \left\{{b},{c}\right\} a \right) \end{aligned}

From this we also get the strictly Pauli algebra identity

\begin{aligned}\left[{\left[{a},{b}\right]},{c}\right] &= {2} \left( a \left\{{b},{c}\right\} - \left\{{b},{c}\right\} a \right) \end{aligned}

But the geometric meaning of this is unfortunately somewhat obfuscated by the notation.

## Generalized dot and wedge product.

The fundamental definitions of dot and wedge products are in terms of grade

\begin{aligned}{\left\langle{{A}}\right\rangle}_{{r}} \cdot {\left\langle{{B}}\right\rangle}_{{s}} = {\left\langle{{AB}}\right\rangle}_{{{{\left\lvert{r-s}\right\rvert}}}} \end{aligned}

\begin{aligned}{\left\langle{{A}}\right\rangle}_{{r}} \wedge {\left\langle{{B}}\right\rangle}_{{s}} = {\left\langle{{AB}}\right\rangle}_{{{r+s}}} \end{aligned}

Use of the trace and Hermitian conjugate split grade selection operations above, we can calculate these for each of the four grades in the Pauli algebra.

There are three dot products consider, vector/vector, bivector/bivector, and trivector/trivector. In each case we want to compute

\begin{aligned}A \cdot B &= \left\langle{{ A}}\right\rangle{ B} \\ &= \frac{1}{{4}}\text{Tr}\left({{ AB + (AB)^{*}}}\right) I \\ &= \frac{1}{{4}}\text{Tr}\left({{ AB + B^{*} A^{*}}}\right) I \end{aligned}

For vectors we have $a^{*} = a$, since the Pauli basis is Hermitian, whereas for bivectors and trivectors we have $a^{*} = -a$. Therefore, in all cases where $A$, and $B$ have equal grades we have

\begin{aligned}A \cdot B &= \left\langle{{A}}\right\rangle{B} I \\ &= \frac{1}{{4}}\text{Tr}\left({{ AB + B A}}\right) I \\ &= \frac{1}{{4}}\text{Tr}{{\left\{{ A},{ B}\right\}}} I \end{aligned}

We have two dot products that produce vectors, bivector/vector, and trivector/bivector, and in each case we need to compute

\begin{aligned}{\left\langle{{ A B}}\right\rangle}_{1} &= \frac{1}{{2}}( AB + (AB)^{*} ) -\frac{1}{{4}}\text{Tr}\left({{ AB + (AB)^{*} }}\right) \end{aligned}

For the bivector/vector dot product we have

\begin{aligned}(Ba)^{*} = -a B \end{aligned}

For bivector $B = i b^k \sigma_k$, and vector $a = a^k \sigma_k$ our symmetric Hermitian sum in coordinates is

\begin{aligned}B a + (Ba)^{*} &= B a - a B \\ &= i b^k \sigma_k a^m \sigma_m - a^m \sigma_m i b^k \sigma_k \end{aligned}

Any $m=k$ terms will vanish, leaving only the bivector terms, which are traceless. We therefore have

\begin{aligned}B \cdot a &= {\left\langle{{ B a}}\right\rangle}_{1} \\ &= \frac{1}{{2}}( B a - a B ) \\ &= \frac{1}{{2}} \left[{B},{a}\right]. \end{aligned}

This result was borrowed without motivation from Clifford algebra in equation (33), and thus not satisfactory in terms of a logically derived sequence.

For a trivector $T$ dotted with bivector $B$ we have

\begin{aligned}(BT)^{*} = (-T)(-B) = TB = BT. \end{aligned}

This is also traceless, and the trivector/bivector dot product is therefore reduced to just

\begin{aligned}B \cdot T &= {\left\langle{{ B T}}\right\rangle}_{1} \\ &= \frac{1}{{2}} \left\{{B},{T}\right\} \\ &= {B}{T} \\ &= T B. \end{aligned}

This is the duality relationship for bivectors. Multiplication by the unit pseudoscalar (or any multiple of it), produces a vector, the dual of the original bivector.

We have two products that produce a grade two term, the vector wedge product, and the vector/trivector dot product. For either case we must compute

\begin{aligned}{\left\langle{{ A B}}\right\rangle}_{2} &= \frac{1}{{2}}( AB - (AB)^{*} ) -\frac{1}{{4}}\text{Tr}\left({{ AB - (AB)^{*} }}\right) \end{aligned} \quad\quad\quad(36)

For a vector $a$, and trivector $T$ we need the antisymmetric Hermitian sum

\begin{aligned}a T - (a T)^{*} &= a T + T a = 2 a T = 2 T a \end{aligned}

This is a pure bivector, and thus traceless, leaving just

\begin{aligned}a \cdot T &= {\left\langle{{ a T}}\right\rangle}_{2} \\ &= a T \\ &= T a \end{aligned}

Again we have the duality relation, pseudoscalar multiplication with a vector produces a bivector, and is equivalent to the dot product of the two.

Now for the wedge product case, with vector $a = a^m \sigma_m$, and $b = b^k \sigma_k$ we must compute

\begin{aligned}a b - (a b)^{*}&= a b - b a \\ &= a^m \sigma_m b^k \sigma_k - b^k \sigma_k a^m \sigma_m \end{aligned}

All the $m = n$ terms vanish, leaving a pure bivector which is traceless, so only the first term of (36) is relevant, and is in this case a commutator

\begin{aligned}a \wedge b &= {\left\langle{{ a b}}\right\rangle}_{2} \\ &= \frac{1}{{2}} \left[{ a},{b}\right] \end{aligned}

There are two ways we can produce a grade three term in the algebra. One is a wedge of a vector with a bivector, and the other is the wedge product of three vectors. The triple wedge product is the grade three term of the product of the three

\begin{aligned}a \wedge b \wedge c&= {\left\langle{{ a b c}}\right\rangle}_{3} \\ &= \frac{1}{{4}} \text{Tr}\left({{ a b c - (a b c)^{*}}}\right) \\ &= \frac{1}{{4}} \text{Tr}\left({{ a b c - c b a}}\right) \end{aligned}

With a split of the $bc$ and $cb$ terms into symmetric and antisymmetric terms we have

\begin{aligned}a b c - c b a &= \frac{1}{{2}} (a \left\{{b},{c}\right\} - \left\{{c},{b}\right\} a)+\frac{1}{{2}} (a \left[{b},{c}\right] - \left[{c},{b}\right] a) \end{aligned}

The symmetric term is diagonal so it commutes (equivalent to scalar commutation with a vector in Clifford algebra), and this therefore vanishes. Writing $B = b \wedge c = \frac{1}{{2}}\left[{b},{c}\right]$, and noting that $\left[{b},{c}\right] = -\left[{c},{b}\right]$ we therefore have

\begin{aligned}a \wedge B&= {\left\langle{{ a B}}\right\rangle}_{3} \\ &= \frac{1}{{4}} \text{Tr}\left({{ a B + B a}}\right) \\ &= \frac{1}{{4}} \text{Tr}{{ \left\{{a},{B}\right\}}} \end{aligned}

In terms of the original three vectors this is

\begin{aligned}a \wedge b \wedge c &= {\left\langle{{ a B}}\right\rangle}_{3} \\ &= \frac{1}{{8}} \text{Tr}{{ \left\{{a},\left[{b},{c}\right]\right\}}}. \end{aligned}

Since this could have been expanded by grouping $ab$ instead of $bc$ we also have

\begin{aligned}a \wedge b \wedge c &= \frac{1}{{8}} \text{Tr}{{ \left\{\left[{a},{b}\right],{c}\right\}}}. \end{aligned}

## Plane projection and rejection.

Projection of a vector onto a plane follows like the vector projection case. In the Pauli notation this is

\begin{aligned}x&= x B \frac{1}{{B}} \\ &= \frac{1}{{2}} \left\{{x},{B}\right\} \frac{1}{{B}} + \frac{1}{{2}} \left[{x},{B}\right] \frac{1}{{B}} \end{aligned}

Here the plane is a bivector, so if vectors $a$, and $b$ are in the plane, the orientation and attitude can be represented
by the commutator

So we have

\begin{aligned}x &= \frac{1}{{2}} \left\{{x},\left[{a},{b}\right]\right\} \frac{1}{{\left[{a},{b}\right]}} + \frac{1}{{2}} \left[{x},{\left[{a},{b}\right]}\right] \frac{1}{{\left[{a},{b}\right]}} \end{aligned}

Of these the second term is our projection onto the plane, while the first is the normal component of the vector.

# Examples.

### Velocity and momentum.

A decomposition of velocity into radial and perpendicular components should be straightforward in the Pauli algebra as it is in the Clifford algebra.

With a radially expressed position vector

\begin{aligned}x = {\left\lvert{x}\right\rvert} \hat{x}, \end{aligned}

velocity can be written by taking derivatives

\begin{aligned}v = x' = {\left\lvert{x}\right\rvert}' \hat{x} + {\left\lvert{x}\right\rvert} \hat{x}' \end{aligned}

or as above in the projection calculation with

\begin{aligned}v&= v \frac{1}{{x}} x \\ &= \frac{1}{{2}}\left\{{v},{\frac{1}{{x}}}\right\} x + \frac{1}{{2}}\left[{v},{\frac{1}{{x}}}\right] x \\ &= \frac{1}{{2}}\left\{{v},{\hat{x}}\right\} \hat{x} + \frac{1}{{2}}\left[{v},{\hat{x}}\right] \hat{x} \end{aligned}

By comparison we have

\begin{aligned}{\left\lvert{x}\right\rvert}' &= \frac{1}{{2}}\left\{{v},{\hat{x}}\right\} \\ \hat{x}' &= \frac{1}{{2 {\left\lvert{x}\right\rvert}}} \left[{v},{\hat{x}}\right] \hat{x} \end{aligned}

In assembled form we have

\begin{aligned}v &= \frac{1}{{2}}\left\{{v},{\hat{x}}\right\} \hat{x} + x \omega \end{aligned}

Here the commutator has been identified with the angular velocity bivector $\omega$

\begin{aligned}\omega &= \frac{1}{{2 x^2}}\left[{x},{v}\right]. \end{aligned}

Similarly, the linear and angular momentum split of a momentum vector is

\begin{aligned}p_\parallel &= \frac{1}{{2}}\left\{{p},{\hat{x}}\right\} \hat{x} \\ p_\perp &= \frac{1}{{2}}\left[{p},{\hat{x}}\right] \hat{x} \end{aligned}

and in vector form

\begin{aligned}p &= \frac{1}{{2}}\left\{{p},{\hat{x}}\right\} \hat{x} + m x \omega \end{aligned}

Writing $J = m x^2$ for the moment of inertia we have for our commutator

\begin{aligned}L = \frac{1}{{2}}\left[{x},{p}\right] = m x^2 \omega = J \omega \end{aligned}

With the identification of the commutator with the angular momentum bivector $L$ we have the total momentum as

\begin{aligned}p &= \frac{1}{{2}}\left\{{p},{\hat{x}}\right\} \hat{x} + \frac{1}{{x}} L \end{aligned}

### Acceleration and force.

Having computed velocity, and its radial split, the next logical thing to try is acceleration.

The acceleration will be

\begin{aligned}a = v' = {\left\lvert{x}\right\rvert}'' \hat{x} + 2 {\left\lvert{x}\right\rvert}' \hat{x}' + {\left\lvert{x}\right\rvert} \hat{x}'' \end{aligned}

We need to compute $\hat{x}''$ to continue, which is

\begin{aligned}\hat{x}''&= \left(\frac{1}{2 {\left\lvert{x}\right\rvert}^3} \left[{v},{x}\right] x \right)' \\ &= \frac{-3}{2 {\left\lvert{x}\right\rvert}^4} {\left\lvert{x}\right\rvert}' \left[{v},{x}\right] x + \frac{1}{2 {\left\lvert{x}\right\rvert}^3} \left[{a},{x}\right] x + \frac{1}{2 {\left\lvert{x}\right\rvert}^3} \left[{v},{x}\right] v \\ &= \frac{-3}{4 {\left\lvert{x}\right\rvert}^5} \left\{{v},{x}\right\} \left[{v},{x}\right] x + \frac{1}{2 {\left\lvert{x}\right\rvert}^3} \left[{a},{x}\right] x + \frac{1}{2 {\left\lvert{x}\right\rvert}^3} \left[{v},{x}\right] v \end{aligned}

Putting things back together is a bit messy, but starting so gives

\begin{aligned}a &= {\left\lvert{x}\right\rvert}'' \hat{x} + 2 \frac{1}{{4 {\left\lvert{x}\right\rvert}^4}} \left\{{v},{x}\right\} \left[{v},{x}\right] x+\frac{-3}{4 {\left\lvert{x}\right\rvert}^4} \left\{{v},{x}\right\} \left[{v},{x}\right] x + \frac{1}{2 {\left\lvert{x}\right\rvert}^2} \left[{a},{x}\right] x + \frac{1}{2 {\left\lvert{x}\right\rvert}^2} \left[{v},{x}\right] v \\ &= {\left\lvert{x}\right\rvert}'' \hat{x} - \frac{1}{4 {\left\lvert{x}\right\rvert}^4} \left\{{v},{x}\right\} \left[{v},{x}\right] x + \frac{1}{2 {\left\lvert{x}\right\rvert}^2} \left[{a},{x}\right] x + \frac{1}{2 {\left\lvert{x}\right\rvert}^2} \left[{v},{x}\right] v \\ &= {\left\lvert{x}\right\rvert}'' \hat{x} + \frac{1}{{4 {\left\lvert{x}\right\rvert}^4}} \left[{v},{x}\right] \left( -\left\{{v},{x}\right\} x + 2 x^2 v \right)+ \frac{1}{2 {\left\lvert{x}\right\rvert}^2} \left[{a},{x}\right] x \end{aligned}

The anticommutator can be eliminated above using

\begin{aligned}v x &= \frac{1}{{2}} \left\{{v},{x}\right\} + \frac{1}{{2}}\left[{v},{x}\right] \\ \implies \\ -\left\{{v},{x}\right\} x + 2 x^2 v &= -(2 v x - \left[{v},{x}\right]) x + 2 x^2 v \\ &= \left[{v},{x}\right] x \end{aligned}

Finally reassembly of the assembly is thus

\begin{aligned}a &= {\left\lvert{x}\right\rvert}'' \hat{x} + \frac{1}{{4 {\left\lvert{x}\right\rvert}^4}} \left[{v},{x}\right]^2 x+ \frac{1}{2 {\left\lvert{x}\right\rvert}^2} \left[{a},{x}\right] x \\ &= {\left\lvert{x}\right\rvert}'' \hat{x} + \omega^2 x+ \frac{1}{2} \left[{a},{x}\right] \frac{1}{{x}} \end{aligned}

The second term is the inwards facing radially directed acceleration, while the last is the rejective component of the acceleration.

It is usual to express this last term as the rate of change of angular momentum (torque). Because $\left[{v},{v}\right] = 0$, we have

\begin{aligned}\frac{d \left[{x},{v}\right]}{dt} = \left[{x},{a}\right] \end{aligned}

So, for constant mass, we can write the torque as

\begin{aligned}\tau&= \frac{d}{dt} \left( \frac{1}{{2}} \left[{x},{p}\right] \right) \\ &= \frac{dL}{dt} \end{aligned}

and finally have for the force

\begin{aligned}F &= m {\left\lvert{x}\right\rvert}'' \hat{x} + m \omega^2 x + \frac{1}{{x}} \frac{dL}{dt} \\ &= m \left({\left\lvert{x}\right\rvert}'' - \frac{{\left\lvert{\omega^2}\right\rvert}}{{\left\lvert{x}\right\rvert}} \right) \hat{x} + \frac{1}{{x}} \frac{dL}{dt} \end{aligned}

# Conclusion.

Although many of the GA references that can be found downplay the Pauli algebra as unnecessarily employing matrices as a basis, I believe this shows that there are some nice computational and logical niceties in the complete formulation of the $\mathbb{R}^{3}$ Clifford algebra in this complex matrix formulation. If nothing else it takes some of the abstraction away, which is good for developing intuition. All of the generalized dot and wedge product relationships are easily derived showing specific examples of the general pattern for the dot and blade product equations which are sometimes supplied as definitions instead of consequences.

Also, the matrix concepts (if presented right which I likely haven’t done) should also be accessible to most anybody out of high school these days since both matrix algebra and complex numbers are covered as basics these days (at least that’s how I recall it from fifteen years back;)

Hopefully, having gone through the exercise of examining all the equivalent constructions will be useful in subsequent Quantum physics study to see how the matrix algebra that is used in that subject is tied to the classical geometrical vector constructions.

Expressions that were scary and mysterious looking like

\begin{aligned}\left[{L_x},{L_y}\right] = i \hbar L_z \end{aligned}

are no longer so bad since some of the geometric meaning that backs this operator expression is now clear (this is a quantization of angular momentum in a specific plane, and encodes the plane orientation as well as the magnitude). Knowing that $\left[{a},{b}\right]$ was an antisymmetric sum, but not realizing the connection between that and the wedge product previously made me wonder “where the hell did the i come from”?

This commutator equation is logically and geometrically a plane operation. It can therefore be expressed with a vector duality relationship employing the $\mathbb{R}^{3}$ unit pseudoscalar $iI = \sigma_1 \sigma_2 \sigma_3$. This is a good nice step towards taking some of the mystery out of the math behind the physics of the subject (which has enough intrinsic mystery without the mathematical language adding to it).

It is unfortunate that QM uses this matrix operator formulation and none of classical physics does. By the time one gets to QM learning an entirely new language is required despite the fact that there are many powerful applications of this algebra in the classical domain, not just for rotations which is recognized (in ([5]) for example where he uses the Pauli algebra to express his rotation quaternions.)

# References

[1] C. Doran and A.N. Lasenby. Geometric algebra for physicists. Cambridge University Press New York, Cambridge, UK, 1st edition, 2003.

[2] D. Hestenes. New Foundations for Classical Mechanics. Kluwer Academic Publishers, 1999.

[3] L. Dorst, D. Fontijne, and S. Mann. Geometric Algebra for Computer Science. Morgan Kaufmann, San Francisco, 2007.

[4] Wikipedia. Pauli matrices — wikipedia, the free encyclopedia [online]. 2009. [Online; accessed 16-June-2009]. http://en.wikipedia.org/w/index.php?title=Pauli_matrices&oldid=29679677%0.

[5] H. Goldstein. Classical mechanics. Cambridge: Addison-Wesley Press, Inc, 1st edition, 1951.