Projecting push forward of a vector

Click For Summary
The discussion revolves around demonstrating that the pushforward of a vector V on manifold M under a diffeomorphism Φ satisfies the equation Φ_*(V) = ⊥(Φ_*V), where ⊥ is a projection operator. The initial proof shows that the additional term in the component form vanishes due to orthogonality with the normal vector. The challenge arises when generalizing this result to (s, 0) tensors, specifically in understanding how the projection interacts with tensor components. Clarification is sought on the concept of orthogonality between vectors and tensors, suggesting that considering tensors as sums of tensor products of basis vectors might aid in comprehension. The discussion emphasizes the need for a deeper understanding of tensor operations in the context of manifold geometry.
Augbrah
Messages
2
Reaction score
0

Homework Statement


Say we have two manifolds N(dim d) and M(dim d-1). Let Φ: M →N be a diffeomorphism where Σ = Φ[M] is hypersurface in N. Let n be unit normal field (say timelike) on Σ and ⊥ projector (in N) defined by:
⊥^a_b = \delta^a_b + n^a n_b

Where acting on (s, 0) tensor projection operator is: ⊥T^{a_1 a_2 ... a_r}=⊥^{a_1} _{b_1}...⊥^{a_s} _{b_s}T^{b_1 ... b_s}

How to show that for a vector V at p on M, the pushforward of V satisfies: Φ_*(V)=⊥(Φ_*V) (I can do that)

And then generalize to (s, 0) tensor: Φ_*(T)=⊥(Φ_*T)

Homework Equations

The Attempt at a Solution


[/B]
In component form we would get: (⊥(Φ_*V))^a = ⊥^a_b (Φ_*V)^b=(\delta^a_b + n^a n_b)(Φ_*V)^b=(Φ_*V)^a + n^an_b(Φ_*V)^b. So remains to show that the last bit vanishes. The last term vanishes since a_b is normal to (Φ_*V)^b, so we're done.

Now to generalize for (s, 0) we have:

(⊥(Φ_*T))^{a_1 a_2 ... a_s} = ^{a_1} _{b_1}...⊥^{a_s} _{b_s}(Φ_*T)^{b_1 ... b_s} = (\delta^{a_1}_{b1} + n^{a_1} n_{b_1}) ... (\delta^{a_s}_{b_s} + n^{a_s} n_{b_s})(Φ_*T)^{b_1 ... b_s}

My problem is that I'm not quite sure what does it mean to act n_{b_s} (Φ_*T)^{b_1 ... b_s}, notion of orthogonality is easily understood for two vectors, but for vector and a tensor I'm not sure. Any ideas much appreciated!
 
Physics news on Phys.org
Would it help to think of the tensor in terms of (a sum of) tensor products of basis vectors?
 

Similar threads

Replies
5
Views
2K
  • · Replies 2 ·
Replies
2
Views
807
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 11 ·
Replies
11
Views
2K
  • · Replies 26 ·
Replies
26
Views
4K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 4 ·
Replies
4
Views
2K