Why Does Frankel Prefer Components on the Right in The Geometry of Physics?

Click For Summary

Discussion Overview

The discussion centers around the preference expressed by Frankel in "The Geometry of Physics" for placing components of vectors to the right of the basis vectors. Participants explore the implications of this notation choice, questioning its advantages and potential confusion compared to more traditional approaches.

Discussion Character

  • Debate/contested
  • Conceptual clarification
  • Technical explanation

Main Points Raised

  • Some participants suggest that Frankel's preference may simply be a matter of taste, arguing that it reads more naturally in certain contexts.
  • Others question whether this notation choice has practical implications in applications such as rotation matrices or Hamilton's principle, seeking clarity on whether it aids understanding or application.
  • One participant emphasizes the importance of being cautious with mathematical notation, noting that the order of components can have deeper significance depending on the context.
  • Another participant points out that while in certain cases, such as with commutative fields, the order may not matter, distinctions can be crucial in other mathematical structures.
  • A later reply mentions that Frankel's reasoning involves achieving a 1x1 result and removing the summation symbol, akin to the Einstein summation convention, while also preserving the notion of vectors as column vectors.
  • Concerns are raised about potential confusion with partial differentiation due to the notation used by Frankel.
  • One participant expresses difficulty in understanding the algebraic context of the discussion and requests a simpler explanation of the implications of the notation choice.
  • Another participant discusses the necessity of distinguishing left and right operations in less well-behaved mathematical objects, suggesting that while vector spaces may allow for some interchangeability, caution is still warranted.

Areas of Agreement / Disagreement

Participants do not reach a consensus on the merits of Frankel's notation preference. Multiple competing views remain regarding its clarity, utility, and potential for confusion.

Contextual Notes

Some participants highlight that the implications of notation can vary significantly depending on the mathematical structures involved, and that the requirement for distinguishing left and right operations may not be universally applicable across all contexts.

JTC
Messages
100
Reaction score
6
Good Day

Early on, in Frankel's text "The Geometry of Physics" (in the introductory note on differential forms, in fact, on page 3) he writes:

"We prefer the last expression with the components to the right of the basis vectors."

Well, I do sort of like this notation and after reading a bit of the text (not easy, struggling, but learning), I am perplexed by one issue: Why?

Why does he prefer to write the expression for a vector with the components on the right of the basis vectors? It is different from the way others do it. Is there something specific that is gained?

(And just to stave off any controversy on the text--yes, I know some do not like this text--I am much less interested on whether his preference is bad, and only interested in WHY he prefers it.)
 
Last edited:
Physics news on Phys.org
JTC said:
Good Day

Early on, in Frankel's text "The Geometry of Physics" (in the introductory note on differential forms, in fact, on page 3) he writes:

"We prefer the last expression with the components to the right of the basis vectors."

Well, I do sort of like this notation and after reading a bit of the text (not easy, struggling, but learning), I am perplexed by one issue: Why?

Why does he prefer to write the expression for a vector with the components on the right of the basis vectors? It is different from the way others do it. Is there something specific that is gained?
I think it's just a matter of taste. Maybe he likes the way it reads: "In direction v go c steps" rather than "Go c steps - huh? where to? - in direction v - see before". E.g. I knew people who actually liked the Polish notation and find the way all others do it confusing.
 
fresh_42 said:
I think it's just a matter of taste. Maybe he likes the way it reads: "In direction v go c steps" rather than "Go c steps - huh? where to? - in direction v - see before". E.g. I knew people who actually liked the Polish notation and find the way all others do it confusing.
Are you suggesting by this that in any application of this notation (e.g.: 1) rotation matrices of frames or 2) Hamilton's principle for linked systems or 3) something I am not even aware of yet) that there is no difference in whether it is before or after? Are you saying that there is no case where the notation makes things easier to understand or apply?

I really do like what you wrote, by the way; and it makes sense. And I suppose I can live with it.

But the flippancy with which Frankel says it, unnerves me just a bit.
 
You must be careful in how you read mathematical notation.

Sometimes its okay to change the order of things and sometimes there's a deeper reason not to.

It varies with the author and the discipline so you need to always be aware of what is what.
 
Formally ##M \cdot R## and ##R \cdot M## with a module ##M## and a ring ##R## are two different objects. E.g. if we chose some matrix rings here to be ##M## and ##R## we can really get something different. (In https://www.physicsforums.com/threads/noncommutative-artinian-rings.885151/#post-5572046 is an example.)

But if we want to write vectors form ##M=V## in coordinates, and have a (commutative) field ##R=F##, we also can swap left and right in all components, so it makes sense to allow ##\alpha \cdot v = v \cdot \alpha\,##. But technically, one has to chose a side.
 
fresh_42 said:
Formally ##M \cdot R## and ##R \cdot M## with a module ##M## and a ring ##R## are two different objects. E.g. if we chose some matrix rings here to be ##M## and ##R## we can really get something different. (In https://www.physicsforums.com/threads/noncommutative-artinian-rings.885151/#post-5572046 is an example.)

But if we want to write vectors form ##M=V## in coordinates, and have a (commutative) field ##R=F##, we also can swap left and right in all components, so it makes sense to allow ##\alpha \cdot v = v \cdot \alpha\,##. But technically, one has to chose a side.

I am afraid I do not understand a sufficient amount of algebra to put your response in context. May I implore you to cut to the chase (forgive me for my facetiousness) and explain in a little more common English, what you are saying?

Are you still suggesting it makes no difference except in so far as we "linguistically" understand: "go in this direction, a certain amount?"
 
I looked at Frankels' notation and don't think its a good idea. He even admits that some may confuse it with partial differentiation.

His reasoning has something to do with getting a 1x1 result and removing the summation symbol kind of like the Einstein summation convention and preserving the notion of v being a column vector.

https://books.google.com/books?id=gXvHCiUlCgUC&pg=PR31&lpg=PR31&dq=frankel+We+prefer+the+last+expression+with+the+components+to+the+right+of+the+basis+vectors.&source=bl&ots=Km9Scq-Vgl&sig=YxycLRxec5l0tObkl1HJbffnMhY&hl=en&sa=X&ved=0ahUKEwif3eyPssDUAhWL64MKHeOwCnAQ6AEILzAB#v=onepage&q=frankel We prefer the last expression with the components to the right of the basis vectors.&f=false
 
Last edited:
JTC said:
I am afraid I do not understand a sufficient amount of algebra to put your response in context. May I implore you to cut to the chase (forgive me for my facetiousness) and explain in a little more common English, what you are saying?
If one has less nicely behaving objects than vector spaces and fields, left and right has to be distinguished, because they might have different properties.
Are you still suggesting it makes no difference except in so far as we "linguistically" understand: "go in this direction, a certain amount?"
Formally, the need for distinction doesn't go away for vector spaces and fields as scalar domains, and I haven't found (on a very quick search) that the requirement ##\alpha \cdot \vec{v} \stackrel{(*)}{=} \vec{v} \cdot \alpha## for a scalar ##\alpha## had been added to the defining properties of vector spaces. However, they don't produce objects with different properties and can be considered isomorphic (equal). But as long as ##(*)## isn't explicitly required, we have to distinguish the two. On the other hand ##(*)## makes kind of sense, since
$$
\alpha \cdot \vec{v} = \alpha \cdot \begin{bmatrix}v_1 \\ v_2\\ \vdots\end{bmatrix} = \begin{bmatrix} \alpha \cdot v_1 \\\alpha \cdot v_2\\ \vdots\end{bmatrix} = \begin{bmatrix}v_1 \cdot \alpha \\ v_2 \cdot \alpha \\ \vdots\end{bmatrix} = \vec{v} \cdot \alpha
$$
and thus we may identify left and right vector spaces. We must not if we regard things lilke matrices as operating objects.
 

Similar threads

  • · Replies 70 ·
3
Replies
70
Views
17K
  • · Replies 8 ·
Replies
8
Views
4K
  • · Replies 15 ·
Replies
15
Views
4K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 16 ·
Replies
16
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 7 ·
Replies
7
Views
4K
Replies
21
Views
4K
  • · Replies 16 ·
Replies
16
Views
7K
  • · Replies 6 ·
Replies
6
Views
3K