Looking for a nice basis for the general linear group

Strilanc
Science Advisor
Messages
612
Reaction score
229
I'm looking for a nice set of basis matrices ##B_{i,j}## that cover the matrices of size ##n \times n## when linear combinations are allowed. The nice property I want them to satisfying is something like ##B_{i,j} \cdot B_{a,b} = B_{i+a, j+b}##, i.e. I want multiplication of two basis matrices to turn into a simple computation inside the indices (using a single simple binary operator).

For example, define ##U^{\oplus x} = \otimes_k^{\lg_2 n} U^{2^i \land k}## where ##\otimes_a^b f(a) = f(0) \otimes f(1) \otimes ... \otimes f(b-1)## is a kronecker product expression, and ##a \land b## returns the number of bit positions where both ##a## and ##b## have a 1-bit in binary. Then an almost-nice basis, for a power-of-2 sized ##n##, is ##B_{i,j} = X^{\oplus i} Z^{\oplus j}##, where ##X## and ##Z## are the usual Pauli matrices. This is the basis of observables that quantum superdense coding is typically done in. Consider:

##B_{i,j} \cdot B_{a,b}##
##= X^{\oplus i} Z^{\oplus j} X^{\oplus a} Z^{\oplus b}##
##= X^{\oplus i} X^{\oplus a} Z^{\oplus j} Z^{\oplus b} (-1)^{a \land j}##
##= X^{\oplus i \oplus a} Z^{\oplus j \oplus b}(-1)^{a \land j}##
##= B_{i \oplus a, j \oplus b} (-1)^{a \land j}##

Which is pretty close to having the operation happen entirely inside the indices with a single operand. There's just that nasty extra sign factor. (The same thing happens if you work from the proper pauli set I, X, Y, Z instead of I, X, Z, and XZ; the factor just gets more complicated.)

Another almost-nice example is the combination of "offset" and "stride" matrices. To make an offset matrix you take the identity matrix then cycle its entries vertically by the given offset. The stride matrices work similarly, except instead of setting every ##i, j## entry satisfying ##i + d = j\pmod{n}## to 1 for any integer ##d## you set every entry satisfying ##i = cj \pmod{n}## for any ##c##.

For example, ##\text{Offset}_1 = \begin{bmatrix} 0 & 0 & 1 \\ 1 & 0 & 0 \\ 0 & 1 & 0 \end{bmatrix}## and ##\text{Stride}_2 = \begin{bmatrix} 1 & 0 & 0 \\ 0 & 0 & 1 \\ 0 & 1 & 0 \end{bmatrix}## when working with ##n = 3##. Offset matrices add their offsets when combined, whereas stride matrices multiply their stride.

If you define ##B_{i,j} = \text{Offset}_i \cdot \text{Stride}_{g^j}## and ##n## is a prime with generator ##g##, then:

##B_{i,j} \cdot B_{a,b}##
##= \text{Offset}_i \cdot \text{Stride}_{g^j} \cdot \text{Offset}_a \cdot \text{Stride}_{g^b}##
##= \text{Offset}_i \cdot \text{Offset}_{a g^j} \cdot \text{Stride}_{g^j} \cdot \text{Stride}_{g^b}##
##= B_{i + a g^j, j+b}##

Which does do all the computation inside the indicates, but doesn't manage to do it with one operator. Also it misses the zero-stride matrices due to the use of the generator.
 
Last edited:
Physics news on Phys.org
Apparently the standard name for what I called "offset" matrices is "shift" matrices. And they're more typically combined with "clock" matrices. That gives you a system where ##B_{i,j} \cdot B_{a,b} = B_{i+a, j+b} \omega^{aj}##, whose extra term is directly analogous to the sign error from the Pauli-matrix-XZ-tensor basis. So still not as simple as desired.
 
Wouldn't this imply you could perform matrix multiplication in ##O(n^2)## essential multiplications, which cannot be done?
 
It depends on how expensive the basis transformation is. If it takes n^3 time to get into the nice multiplication basis, or if the basis has n^1.5 elements due to necessary redundancy, then it wouldn't help much with multiplication. The basis examples I gave are in fact cheap to get into though, and applying it to multiplication is what I had in mind.
 
##\textbf{Exercise 10}:## I came across the following solution online: Questions: 1. When the author states in "that ring (not sure if he is referring to ##R## or ##R/\mathfrak{p}##, but I am guessing the later) ##x_n x_{n+1}=0## for all odd $n$ and ##x_{n+1}## is invertible, so that ##x_n=0##" 2. How does ##x_nx_{n+1}=0## implies that ##x_{n+1}## is invertible and ##x_n=0##. I mean if the quotient ring ##R/\mathfrak{p}## is an integral domain, and ##x_{n+1}## is invertible then...
The following are taken from the two sources, 1) from this online page and the book An Introduction to Module Theory by: Ibrahim Assem, Flavio U. Coelho. In the Abelian Categories chapter in the module theory text on page 157, right after presenting IV.2.21 Definition, the authors states "Image and coimage may or may not exist, but if they do, then they are unique up to isomorphism (because so are kernels and cokernels). Also in the reference url page above, the authors present two...
When decomposing a representation ##\rho## of a finite group ##G## into irreducible representations, we can find the number of times the representation contains a particular irrep ##\rho_0## through the character inner product $$ \langle \chi, \chi_0\rangle = \frac{1}{|G|} \sum_{g\in G} \chi(g) \chi_0(g)^*$$ where ##\chi## and ##\chi_0## are the characters of ##\rho## and ##\rho_0##, respectively. Since all group elements in the same conjugacy class have the same characters, this may be...

Similar threads

Back
Top