Strange index notation for linear transformation matrix

  • Context: Undergrad 
  • Thread starter Thread starter sphyrch
  • Start date Start date
  • Tags Tags
    Index Linear Notation
Click For Summary

Discussion Overview

The discussion revolves around the notation used for linear transformation matrices in the context of basis transformations in vector spaces, particularly in relation to dual bases and tensor representations. Participants explore the implications of staggered index notation and its relevance to tensor types and transformations.

Discussion Character

  • Technical explanation
  • Conceptual clarification
  • Debate/contested

Main Points Raised

  • One participant notes that the staggered index notation in the transformation ##e'_j=A^i_{\ \ j}e_i## is meant to distinguish between types of tensors and their operations, suggesting that this notation is crucial for clarity in tensor algebra.
  • Another participant proposes that the dual basis transformation can be represented as $$\epsilon'^j=(\tilde A^{-1})_i^{\ \ \ j}\epsilon^i$$ and questions the reasoning behind the staggered notation for ##(\tilde A^{-1})##, considering it as a tensor in the context of dual spaces.
  • A different participant expresses confusion about why the indices are staggered in the notation for the transpose of the matrix, indicating a lack of clarity on the conventions used in the literature.
  • One participant suggests that the arrangement of indices is a matter of convention, referencing different authors' preferences for index placement in tensor notation, and emphasizes the importance of consistency in notation.
  • Another participant cautions against conflating matrix algebra with tensor notation, suggesting that while the transformation matrix can be treated similarly, it is essential to recognize the differences in their mathematical frameworks.

Areas of Agreement / Disagreement

Participants express differing views on the conventions for index notation, with some favoring specific arrangements while others highlight the variability in approaches among different authors. The discussion remains unresolved regarding the best practices for notation in this context.

Contextual Notes

Participants acknowledge that there are multiple conventions for index notation in tensor calculus, which can lead to confusion. The discussion reflects the complexity of representing transformations and the need for clarity in mathematical communication.

sphyrch
Messages
37
Reaction score
9
I'm reading Liang's book on General Relativity and Differential Geometry, and came across this part:

If there is a basis transformation ##e'_j=A^i_{\ \ j}e_i## in a vector space ##V## and the (non-degenerate) matrix constituted by elements ##A^i_{\ \ j}## is denoted by ##A##, then the corresponding dual basis transformation is $$\epsilon'^j=(\tilde A^{-1})_i^{\ \ \ j}\epsilon^i$$ where ##\tilde A=A^T## (transpose)

Remark: Here we write the matrix elements as ##A^i_{\ \ j}##. The reason for distinguishing the upper and lower indices is to distinguish summation and to distinguish the type of a tensor. However, what is important in the matrix operation is just differentiating the left and right indices. Therefore, if you want, you may change all upper indices to lower indices for now; for instance the last equation can be written as ##\epsilon'_j=(\tilde A^{-1})_{ij}\epsilon_i##.

I just want to have a crystal clear understanding of why this notation is chosen. Basis transformation would be an automorphism from ##V## to ##V##, and there's a result saying that the set of automorphisms is isomorphic to ##T_1^1(V)##. So ##A## can be identified with a tensor that eats a covector and a vector => so its index representation (i.e. element indexing) is to be identified with some ##(1,1)## tensor ##T##, i.e. ##T^i_{\ \ j}=T(\epsilon^i,e_j)##. So we also represent ##A## by ##A^i_{\ \ j}##.

But I'm confused why we're writing staggered index notation for ##(\tilde A^{-1})## as ##(\tilde A^{-1})_i^{\ \ \ j}## - is there some concrete reason? Should I imagine ##(\tilde A^{-1})## as some automorphism from ##V^*## to ##V^*##, and using similar logic as before, say that it can be identified with some tensor ##\tilde T\in T^1_1(V^*)##? In that case, elements of ##\tilde T## are represented as ##\tilde T(e'_i,\epsilon'^j)=\tilde T_i^{\ \ \ j}##, so we represent ##(\tilde A^{-1})## by ##(\tilde A^{-1})_i^{\ \ \ j}##?

For the dual basis transformation, why can't I just write ##\epsilon'^j=A^j_{\ \ \ k}\epsilon^k##, so that ##A^i_{\ \ \ j}A^j_{\ \ \ k}=\delta^i_{\ \ \ k}##. OR why can't I write ##\epsilon'^j=A^j_{\ \ \ k}\epsilon^k## and ##\epsilon'_j=A_j^{\ \ \ i}e_i##, so that ##A_j^{\ \ \ i}A^j_{\ \ \ k}=\delta^i_{\ \ \ k}##?
 
Physics news on Phys.org
If you have a vector ##V^i## and a metric ##g_{ij}## then you have the dual vector ##\omega_i=g_{ij}V^j##. If you also have a vector basis transformation matrix ##A^i{}_j## then$$\begin{eqnarray*}
V'^i&=&A^i{}_jV^j\\
g'_{ij}V'^j&=&g'_{ik}A^k{}_jV^j\\
\omega'_i&=&g'_{ik}A^k{}_jg^{jl}\omega_l
\end{eqnarray*}$$We would like the ##g'Ag## term on the right to be our transformation matrix for covectors. Obviously we can just multiply out the components to get a single matrix, and it makes sense to follow the rules for tensors and say that the metric and its inverse lowered and raised an index. Hence the transformation matrix should have lower and upper indices.
 
Thanks for the response! But I'm sorry @Ibix I couldn't follow - I think I did not frame my question very well. My doubt isn't why one index is up and the other is down - it's why the indices are staggered in that particular way. There are three paragraphs in my OP, first para I'm pretty sure about (though you can correct if I'm wrong). So that explains why indices are staggered the way they are in ##e'_j=A^i_{\ \ \ j}e_i##.

Regarding 2nd paragraph, that's a hand-wavy explanation I've cooked up that I'm not at all sure about. I don't even know why the transpose of ##A^i_{\ \ \ j}## is written as ##\tilde A^{\ \ \ i}_j##. See here (an excerpt from the book) for an example - 3rd equality first term:
1714987481876.png

And my question in the final paragraph still stands..
 
If I understand the question in the last paragraph, the reason why the indices are laid out the way they are is a convention. For example, Carroll's lecture notes use this one but note that Schutz lays out coordinate transforms with northwest/southeast indices always.

I don't have strong feelings either way (except to wish that we could get all physicists together and agree a set of conventions for once...), but I have a weak preference for the standard in this thread. If you had a tensor ##T^i{}_j## and you applied the metric and inverse metric ##g_{ik}T^k{}_lg^{lj}## you would write the result as ##T_i{}^j##, wouldn't you? ##A## is a matrix not a tensor, but you can still multiply out the components of ##g'_{ik}A^k{}_lg^{lj}## and it would make sense to place the indices in a consistent way.

I have to say that matrix algebra and tensor notation aren't quite the same thing and you can mislead yourself if you try to see too many parallels. I don't think you "have to" notate a transpose in any particular way. It doesn't even make sense to talk about the transpose of a mixed-index tensor. I think it works for the transformation matrix, but I'm not sure that I'd think of it that way.
 
Ibix said:
For example, Carroll's lecture notes use this one but note that Schutz lays out coordinate transforms with northwest/southeast indices always
Note that Carroll'book, instead, uses only northwest/southeast indices.
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 12 ·
Replies
12
Views
2K
  • · Replies 1 ·
Replies
1
Views
944
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 4 ·
Replies
4
Views
4K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 7 ·
Replies
7
Views
996