- 10,876
- 423
Here's my take on that part of the proof. I think I've made it to eq. (B3), but like you (if I understand you correctly), I have ##x_0## where they have ##x##. I'll write t,s instead of λ,λ' because it's easier to type, and I'll write u instead of v' because I'm going to use primes for derivatives, so I don't want any other primes. I will denote the map that takes straight lines to straight lines by ##\Lambda##, because that's a fairly common notation for a change of coordinates, and because seeing it written as x' really irritates me.Erland said:Strangerep, I must admire your patience. Yes, I suppose one must spend months if one should get a chance to understand this proof by Guo et al. A proof that to me seems to be utter gibberish. Even if their reasoning probably is correct, they have utterly failed to communicate it in an intelligible way.
But since you claim you now understand it. I keep asking you about it. I hope that's okay...What do you mean by "pick a parametrization"? How is this picking administered? Surely, such parametrizations cannot be picked in a completely arbitrary manner, not even depending continuously upon the lines (or their positions)?
The only way I can understand this is to consider a map from lines to lines, but not lines as point sets, but as parametrized lines. If (x0,v) determines a parametization x=x0+λv of a line, this is mapped to M(x0,v)=(y0,w) where y0=T(x0) and w=T(x0+v)-T(x0), where T is the coordinate transformation.
But even so, f(x,v) should be a function of x0, v and λ, not of x. And I don't understand how they can claim that f depends linearly upon v. This seems outright false, since we have the factors vivj, which is a quadratic expression in v, not a linear one. And then they deduce an equation (B3) in a way that I don't understand either.
So, there is not much I understand in this proof.![]()
Let x be an arbitrary vector. Let v be an arbitrary non-zero vector. The map ##t\mapsto x+tv## (with domain ℝ) is a straight line. (Note that my x is their x0). By assumption, ##\Lambda## takes this to a straight line. So ##\Lambda(x)## is on that line, and for all t in ℝ, ##\Lambda(x+tv)## is on that line too. This implies that there's a non-zero vector u (in the codomain of ##\Lambda##) such that for each t, there's an s such that ##\Lambda(x+tv)=\Lambda(x)+su##.
Since we're dealing with a finite-dimensional vector space, let's define a norm on it and require u to be a unit vector. Now the number s is completely determined by the properties of ##\Lambda## along the straight line ##t\mapsto x+tv##, which is completely determined by x and v. It would therefore be appropriate to write the last term of ##\Lambda(x)+su## as s(x,v,t)u(x,v), but that would clutter the notation, so I will just write s(t)u. We will have to remember that they also depend on x and v. I will write the partial derivative of s with respect to t as s'. So, for all t, we have
$$\Lambda(x+tv)=\Lambda(x)+s(t)u.\qquad (1)$$ Now take the ith component of (1) and Taylor expand both sides around t=0. I will use the notation ##{}_{,j}## for the jth partial derivative. The first-order terms must be equal:
$$t\Lambda^i{}_{,j}(x)v^j=ts'(0)u.$$ This implies that
$$u=\frac{\Lambda^i{}_{,j}(x) v^j}{s'(0)}.$$ Now differentiate both sides of the ith component of (1) twice with respect to t, and then set t=0.
$$\Lambda^i{}_{,jk}(x)v^jv^k =s''(0)u=\frac{s''(0)}{s'(0)}\Lambda^i{}_{,j}(x) v^j.\qquad(2)$$ Now it's time to remember that s(t) really means s(x,v,t). The value of s''(0)/s'(0) depends on x and v, and is fully determined by the values of those two variables. So there's a function f such that ##f(x,v)=s''(0)/s'(0)##. Let's postpone the discussion of whether f must be linear in the second variable, and first consider what happens if it is linear in the second variable. Then we can write ##f(x,v)=v^i f_{,i}(x,0)=2f_{i}(x)v^i##, where I have defined ##f_i## by ##f_i(x)=f_{,i}(x,0)/2##. The reason for the factor of 2 will be obvious below. Now we can write (2) as
\begin{align}
\Lambda^i{}_{,jk}(x)v^jv^k &=2f_k(x)\Lambda^i{}_{,j}(x) v^j v^k\\
&=f_k(x)\Lambda^i{}_{,j}(x) v^j v^k +f_k(x)\Lambda^i{}_{,j}(x) v^j v^k\\
&=f_k(x)\Lambda^i{}_{,j}(x) v^j v^k +f_j(x)\Lambda^i{}_{,k}(x) v^k v^j\\
&=\big(f_k(x)\Lambda^i{}_{,j}(x)+f_j(x)\Lambda^i{}_{,k}(x)\big)v^k v^j.\qquad (3)
\end{align} All I did to get the third line from the second was to swap the dummy indices j and k in the second term. Since (3) holds for all x and all v≠0, it implies that
$$\Lambda^i{}_{,jk}(x)=f_k(x)\Lambda^i{}_{,j}(x)+f_j(x)\Lambda^i{}_{,k}(x).\qquad (4)$$ This is my version of their (B3). Since my x is their x0, it's not exactly the same. The fact that they have x (i.e. my x+tv) in the final result suggests that they didn't set t=0 like I did. So I think their result is equivalent to mine even though it looks slightly different.Let's get back to the linearity of f in the second variable. I don't have a perfect argument for it yet, but I'm fairly sure that it can be proved using arguments similar to this (even though this one doesn't quite go all the way): (2) is an equality of the form
$$v^T M v= g(v)m^Tv,$$ where M is an n×n matrix and m is an n×1 matrix (like v). The equality is supposed to hold for all v. For all ##a\in\mathbb R##, we have
$$g(av)m^Tv =\frac{g(av)Mg(av)}{a} =\frac{1}{a}(av)^TM(av) =av^TMv =ag(v)m^Tv.$$ So at least we have ##g(av)=ag(v)## for all v such that ##m^Tv\neq 0##.
Last edited: