Manipulating vector-matrices combinations

In summary, the author is trying to understand differential geometry in terms of general relativity. However, they are unsure of some fundamental maths related to the metric. They ask a summarizer for help and clarification, but the summarizer is not sure how matrix multiplication explains the process. They would appreciate any clarification offered.
  • #1
mysearch
Gold Member
526
0
Hi,
I am trying to understand some of the basics of differential geometry in respect to general relativity. However, I not sure that I understand what seems to be a fairly fundamental bit of maths connected with the generic metric given in [1]. For simplicity, the notation has been reduced to 2-dimensional (x,y) coordinates:

[1] [tex]ds^2=g_{xy} dx^x dx^y[/tex]

While I understand the geometry by which the separation (ds) can be expanded to the form in [2], I am unsure how matrix multiplication explains this process.

[2] [tex]ds^2 = g_{11} xx + g_{12} xy + g_{21} yx +g_{22} yy [/tex]

In the simplest case, [2] collapses to Pythagoras via the values assigned to (g). For example, in 2D space, I was assuming (g=1,0,0,1) to be a 2x2 matrix, while (dx dy) might be described as 2x1 matrices that represent the component vectors of (ds) along the orthogonal axes (x,y), e.g.(x=3,0) and (y=0,4).

[3] [tex]ds^2 = \left(\begin{array}{cc}1&0\\0&1\end{array}\right) \left(\begin{array}{cc}3\\0\end{array}\right) \left(\begin{array}{cc}0\\4\end{array}\right) [/tex]

While this logic, which is likely to be flawed, leads to the form in [3], I am not sure how multiplying the three matrices with the dimensions above leads to the form suggested by [2]. Therefore, I would much appreciate any clarification on offer. Thanks
 
Physics news on Phys.org
  • #2
mysearch said:
Hi,
For simplicity, the notation has been reduced to 2-dimensional (x,y) coordinates:

[1] [tex]ds^2=g_{xy} dx^x dx^y[/tex]
In that case, don't you mean

[tex]ds^2=g_{xx}(dx^x)^2+g_{yy}(dx^y)^2[/tex]

or

[tex]ds^2=g_{xx}dx^2+g_{yy}dy^2[/tex]

?

What you wrote looks like only one of the n2 terms of [itex]g_{ij}dx^i dx^j[/itex].

I don't quite understand what you're saying about matrix multiplication, so I'll just write a few comments about the expression [itex]ds^2=g_{ij}dx^i dx^j[/itex].

Do you understand this equality?

[tex]g=g_{ij} dx^i\otimes dx^j[/tex]

If we define

[tex]dx^i dx^j=\frac{1}{2}(dx^i\otimes dx^j+dx^j\otimes dx^i)[/tex]

and use this definition to rewrite the right-hand side of that equality, we get

[tex]g=g_{ij} dx^i\otimes dx^j=\frac{1}{2}(g_{ij}dx^i\otimes dx^j+g_{ji}dx^j\otimes dx^i)=g_{ij}dx^i dx^j[/tex]

So if we define ds2=g, we have made the equality [itex]ds^2=g_{ij}dx^idx^j[/itex] come true.

Edit: Also, consider this: If C is a curve and \dot C its velocity vector field, then

[tex]g(\dot C,\dot C)=g_{ij}\dot C^i\dot C^j[/tex]

If we write the components of the velocity vector field as dxi/dt and think of dxi as an approximation of the change of the ith coordinate of C(t) when t changes by dt, we can write

[tex]g(\dot C,\dot C)dt^2=g_{ij}\frac{dx^i}{dt}\frac{dx^j}{dt}dt^2=g_{ij}dx^i dx^j[/tex]

This suggests that the proper way to define the arc length L(C) of a curve C:[a,b]→M in a Riemannian manifold M is

[tex]L(C)=\int_a^b g_{C(t)}(\dot C(t),\dot C(t))dt[/tex]

and this is how arc length is defined.
 
Last edited:
  • #3
Fredrik,
Many thanks for your comments. I appreciate that I am probably not interpreting or using some of the notation correctly as the subject of differential geometry and tensors is quite new to me. So, at this stage, I am simply trying to establish some basic understanding of metrics and the metric tensors that underpin GR by plodding through the use of superscripts and subscripts in connection with contravariant and covariant coordinates. This led me into a discussion of orthogonal and oblique coordinates systems. Using just standard geometry and limiting this discussion to 2D, these systems seem to lead to the equation of the form:

[1] [tex]ds^2 = g_{11}x^1 x^1 + g_{12} x^1 x^2 + g_{21} x^2 x^1 +g_{22} x^2 x^2 [/tex]

While some of these terms collapse to zero in Pythagoras' theorem, they are present in other metrics. By and large this made sense to me. However, many texts then seem to describe [1] in terms of the following equation, which I have copied from your post:

[2] [tex]ds^2=g_{ij}dx^i dx^j[/tex]

Given that both equate to (ds) I assume that the RHS of [1] is an expanded form of the RHS of [2] in the case of a 2D solution:

[3] [tex]g_{ij}dx^i dx^j \equiv g_{11}x^1 x^1 + g_{12} x^1 x^2 + g_{21} x^2 x^1 +g_{22} x^2 x^2[/tex]

However, I don’t understand what rules drive this expansion or whether it means something completely different.
Fredrik said:
In that case, don't you mean
[tex]ds^2=g_{xx}(dx^x)^2+g_{yy}(dx^y)^2[/tex]
or
[tex]ds^2=g_{xx}dx^2+g_{yy}dy^2[/tex]?
What you wrote looks like only one of the n2 terms of [itex]g_{ij}dx^i dx^j[/itex].
Quite possibly, but where did the power of 2 appear from based on [2] above? Does it fall out of the summation notation suggested below in [4]?

[4] [tex]ds^2=g_{ij}dx^i dx^j = ds^2=g_{ij} \sum dx^i \sum dx^j [/tex]
Fredrik said:
Do you understand this equality?
[tex]g=g_{ij} dx^i\otimes dx^j[/tex]
No, I have never used this notation, but it seems to be used in connection with tensor multiplication (?). However, if possible, I would really like to try to resolve my initial misunderstandings first, before proceed into to yet more unknowns:redface:
 
  • #4
mysearch,

You're mostly on track. Starting with your [2]

[tex]ds^2 = g_{11}x^1 x^1 + g_{12} x^1 x^2 + g_{21} x^2 x^1 +g_{22} x^2 x^2 [/tex],

we can write this more compactly as the double sum

[tex]ds^2 = \sum_i \sum_j g_{ij}x^i x^j [/tex],

and then even more compactly by just inferring the sums from the repeated indices

[tex]ds^2=g_{ij}dx^i dx^j[/tex].

The connection with matrix notation is

[tex]
\left[ \begin{array}{cc} x^1 & x^2 \end{array} \right]
\left[ \begin{array}{cc} g_{11} & g_{12} \\ g_{21} & g_{22} \end{array} \right]
\left[ \begin{array}{c} x^1 \\ x^2 \end{array} \right]
= g_{11}x^1 x^1 + g_{12} x^1 x^2 + g_{21} x^2 x^1 +g_{22} x^2 x^2
[/tex]

but I wouldn't say it explains things. It is just another way to write it.
 
  • #5
Uncorrelated said:
[1] [tex]ds^2 = g_{11}x^1 x^1 + g_{12} x^1 x^2 + g_{21} x^2 x^1 +g_{22} x^2 x^2 [/tex],

[2] [tex]ds^2 = \sum_i \sum_j g_{ij}x^i x^j [/tex],

[3] [tex]ds^2=g_{ij}dx^i dx^j[/tex].
Many thanks for the excellent summary in post #4. I had managed to work through most of the ever-increasing shorthand styles of notation that finishes up in the form of [3] above. However, I had lost my way when trying to re-interpret [3] back towards [1] via matrix multiplication.
Uncorrelated said:
[4] [tex]
\left[ \begin{array}{cc} x^1 & x^2 \end{array} \right]
\left[ \begin{array}{cc} g_{11} & g_{12} \\ g_{21} & g_{22} \end{array} \right]
\left[ \begin{array}{c} x^1 \\ x^2 \end{array} \right]
= g_{11}x^1 x^1 + g_{12} x^1 x^2 + g_{21} x^2 x^1 +g_{22} x^2 x^2
[/tex]
This is the step I had mis-understood. I had seen a vector described as a vertical [2x1] matrix and had assumed that this orientation was a requirement for both of vectors implied in [2]. I had also come across the rule that 2 matrices can only be multiplied together if the number of columns in the first is the same as the number of rows in the second. If this is the case, would the ordering of the vector matrices be important?

[tex] \left[2x1 \right] \left[1x2 \right] = \left[2x2 \right] [/tex]

[tex]\left[ \begin{array}{c} x^1 \\ x^2 \end{array} \right]
\left[ \begin{array}{cc} x^1 & x^2 \end{array} \right]
= x^1 x^1 + x^2 x^1 + x^1 x^2 + x^2 x^2[/tex]

Presumably, if this is correct, the resulting [2x2] matrix would then be multiplied by (g), another [2x2] matrix.
Uncorrelated said:
but I wouldn't say it explains things. It is just another way to write it.

I couldn’t agree more. In many respect, the form of equation [2] seems to be the important aspect, which depends on the coordinate system, i.e. number of dimensions and the nature of space/spacetime assumed. While I appreciate that differential geometry may be a very useful mathematical tool, it appears to lead to so much abstracted notation it is quite easy to lose sight of any physical reality. Well, that’s my excuse anyway:smile:
 
Last edited:
  • #6
Yes, when you write things in matrix form the order and whether vectors are row or column matters and you get different kinds of multiplication. The basic one is the inner product which results in a number

[tex]
\left[ \begin{array}{cc} x^1 & x^2 \end{array} \right]
\left[ \begin{array}{c} y^1 \\ y^2 \end{array} \right]
= x^1 y^1 + x^2 y^2
[/tex]
.

Then there is the outer product (I'm not sure everyone uses the same convention for this) which results in a matrix

[tex]
\left[ \begin{array}{c} x^1 \\ x^2 \end{array} \right]
\left[ \begin{array}{cc} y^1 & y^2 \end{array} \right]
= \left[ \begin{array}{cc} x^1 y^1 & x^1 y^2 \\ x^2 y^1 & x^2 y^2 \end{array}\right]
[/tex]
.

If you think of the matrix as a stack of row vectors or a bunch of column vectors lined up

[tex]
\left[ \begin{array}{cc} x^{11} & x^{12} \\ x^{21} & x^{22} \end{array}\right]
=
\begin{array}{c}
\left[ \begin{array}{cc} x^{11} & x^{12} \end{array} \right] \\
\left[ \begin{array}{cc} x^{21} & x^{22} \end{array} \right]
\end{array}
=
\begin{array}{cc}
\left[ \begin{array}{c} x^{11} \\ x^{21} \end{array} \right] &
\left[ \begin{array}{c} x^{12} \\ x^{22} \end{array} \right]
\end{array}
[/tex]

then everything is just repeated application of these two multiplications.

--
Uncorrelated

mysearch said:
This is the step I had mis-understood. I had seen a vector described as a vertical [2x1] matrix and had assumed that this orientation was a requirement for both of vectors implied in [2]. I had also come across the rule that 2 matrices can only be multiplied together if the number of columns in the first is the same as the number of rows in the second. If this is the case, would the ordering of the vector matrices be important?

[tex] \left[2x1 \right] \left[1x2 \right] = \left[2x2 \right] [/tex]

[tex]\left[ \begin{array}{c} x^1 \\ x^2 \end{array} \right]
\left[ \begin{array}{cc} x^1 & x^2 \end{array} \right]
= x^1 x^1 + x^2 x^1 + x^1 x^2 + x^2 x^2[/tex]

Presumably, if this is correct, the resulting [2x2] matrix would then be multiplied by (g), another [2x2] matrix.
 
  • #7
mysearch said:
However, I had lost my way when trying to re-interpret [3] back towards [1] via matrix multiplication.
Is this the definition of matrix multiplication you're using: (AB)ij=AikAkj? (It should be).

mysearch said:
I had also come across the rule that 2 matrices can only be multiplied together if the number of columns in the first is the same as the number of rows in the second.
Note that this is implied by the definition (AB)ij=AikAkj.

mysearch said:
Using just standard geometry and limiting this discussion to 2D, these systems seem to lead to the equation of the form:

[1] [tex]ds^2 = g_{11}x^1 x^1 + g_{12} x^1 x^2 + g_{21} x^2 x^1 +g_{22} x^2 x^2 [/tex]
You shouldn't use the d(something) notation on the left if you're not going to do it on the right. But you're right about how to apply the summation convention. The right-hand side is [itex]g_{ij}x^iy^j[/itex].

mysearch said:
Quite possibly, but where did the power of 2 appear from based on [2] above? Does it fall out of the summation notation suggested below in [4]?

[4] [tex]ds^2=g_{ij}dx^i dx^j = ds^2=g_{ij} \sum dx^i \sum dx^j [/tex]
[itex]ds^2[/itex] is defined by [itex]ds^2=g_{ij}dx^idx^j[/itex].

mysearch said:
No, I have never used this notation, but it seems to be used in connection with tensor multiplication (?). However, if possible, I would really like to try to resolve my initial misunderstandings first, before proceed into to yet more unknowns:redface:
In that case, it's probably impossible to discuss the metric tensor. The best we can do is to discuss norms defined from inner products defined from matrix multiplication. Let g be an arbitrary non-singular symmetric 2×2 matrix, and define the inner product on the vector space of 2×1 matrices by

[tex]\langle x,y\rangle=x^Tgy[/tex]

and the norm by

[tex]\|x\|=\sqrt{\langle x,x\rangle}[/tex]

(You should verify that the above defines an inner product and a norm). These definitions imply

[tex]x^2=\langle x,x\rangle=x^Tgx=x_ig_{ij}x_j=g_{ij}x_i x_j[/tex]

If g=I, we get

[tex]x^2=\delta_{ij}x_ix_j=(x_1)^2+(x_2)^2[/tex]
 
  • #8
Fredrik & Uncorrelated,
Really appreciate all the help as some of the pieces now make more sense. However, I will try to work through all the information provided and read a little more into the maths and notation behind the metric tensor. Clearly, all the ideas contained in differential geometry is going to take a bit more work on my part.:cry:

Anyway, many thanks and a Merry Xmas.:smile:
 

1. What is a vector-matrix combination?

A vector-matrix combination is a mathematical operation where a vector (a one-dimensional array of numbers) is multiplied by a matrix (a two-dimensional array of numbers). The result of this operation is a new vector with updated values based on the matrix multiplication.

2. How do you manipulate vector-matrix combinations?

Vector-matrix combinations can be manipulated in several ways, including scalar multiplication, addition and subtraction, and matrix multiplication. Scalar multiplication involves multiplying each element of the vector by a single scalar value. Addition and subtraction can be done between two vectors or between a vector and a matrix, following certain rules. Matrix multiplication is a more complex operation that involves multiplying two matrices to create a new matrix.

3. What are the applications of manipulating vector-matrix combinations?

Vector-matrix combinations are used in various fields, including physics, engineering, economics, and computer science. They are especially useful in data analysis, machine learning, and computer graphics. For example, in computer graphics, vector-matrix combinations are used to transform 3D objects on a 2D screen.

4. Is it possible to manipulate vector-matrix combinations in different dimensions?

Yes, it is possible to manipulate vector-matrix combinations in different dimensions. For example, a 2D vector can be multiplied by a 3D matrix, resulting in a new 2D vector. However, the dimensions of the vector and matrix must be compatible for the operation to be performed.

5. How can I learn more about manipulating vector-matrix combinations?

There are many online resources and textbooks available to learn more about manipulating vector-matrix combinations. Some recommended topics to study include matrix operations, linear algebra, and vector calculus. Additionally, practicing with different examples and problems can help improve understanding and proficiency in this subject.

Similar threads

  • Linear and Abstract Algebra
Replies
3
Views
2K
  • Advanced Physics Homework Help
Replies
19
Views
1K
  • Special and General Relativity
Replies
5
Views
1K
  • Differential Geometry
Replies
7
Views
2K
  • Differential Geometry
Replies
4
Views
3K
  • Advanced Physics Homework Help
Replies
19
Views
943
Replies
82
Views
2K
  • MATLAB, Maple, Mathematica, LaTeX
Replies
3
Views
550
Replies
13
Views
504
  • Linear and Abstract Algebra
Replies
7
Views
928
Back
Top