Sciencemaster
- 129
- 20
- TL;DR Summary
- I understand Einstein summation on its own, but when I try to connect tensor notation directly to linear algebra objects, I get tripped up by how vectors and tensors should be represented. How should I properly think about this connection?
I've been reviewing some introductory tensor stuff, and I've come to the realization that some of the things tensors do confuse me. For example, the notes I'm reading say that the invariant interval is both ##S=\eta_{\mu\nu} x^\mu x^nu## and ##S=x^T \eta x##. Both of which are totally fine on their own, the former sums over each vector with the proper sign from the metric, and the second gives the same scalar value via matrix multiplication. My issue is with equating the two. The way the former is written, I think of the first one as the metric matrix times two coordinate column vectors (because of their upper indices). In terms of linear algebra, not only would this be a column vector rather than a scalar, but neither of the vectors would be transposed since both ##x^\mu## and ##x^\nu## are both after the matrix and have their indices in the same place (i.e. on equal footing).
Upon further reflection, I realized that just the idea that ##a_{\mu\nu} b^\mu## would take a column vector (upper index) and matrix multiply it to produce a row vector (lower index) seems strange with linear algebra sensibilities. In the past, I have worked both with linear algebra and tensors just fine, Einstein summation notation makes sense in itself and I've manipulated my fair share of tensor equations. However, going back and trying to conceptually connect Tensor math with the presented framework of linear algebra, I'm having trouble with the connection between the two areas of math. Since (0,1)-tensors have been presented as row vectors thus far, I'm tempted to visualize a (0,2) tensor as a collection of row vectors rather than a "normal" matrix given how it interacts with (1,0)-tensors (summing with two column vectors to create a scalar or summing over one column vector and leaving one row vector as the output). The overall point is, I'm struggling with the connection between linear algebra and basic tensor math.
Upon further reflection, I realized that just the idea that ##a_{\mu\nu} b^\mu## would take a column vector (upper index) and matrix multiply it to produce a row vector (lower index) seems strange with linear algebra sensibilities. In the past, I have worked both with linear algebra and tensors just fine, Einstein summation notation makes sense in itself and I've manipulated my fair share of tensor equations. However, going back and trying to conceptually connect Tensor math with the presented framework of linear algebra, I'm having trouble with the connection between the two areas of math. Since (0,1)-tensors have been presented as row vectors thus far, I'm tempted to visualize a (0,2) tensor as a collection of row vectors rather than a "normal" matrix given how it interacts with (1,0)-tensors (summing with two column vectors to create a scalar or summing over one column vector and leaving one row vector as the output). The overall point is, I'm struggling with the connection between linear algebra and basic tensor math.