Cauchy Schwarz equality implies parallel

  • Context: Graduate 
  • Thread starter Thread starter Bipolarity
  • Start date Start date
  • Tags Tags
    Cauchy Parallel
Click For Summary

Discussion Overview

The discussion revolves around proving that if the inner product of two non-zero vectors in an inner product space over the complex numbers equals the product of their norms, then one vector is a scalar multiple of the other. This topic intersects linear algebra and its applications in machine learning, particularly in the context of Support Vector Machines.

Discussion Character

  • Exploratory
  • Technical explanation
  • Mathematical reasoning

Main Points Raised

  • One participant seeks to prove that if ## \langle v1 , v2 \rangle = ||v1|| \cdot ||v2|| ##, then ##v1## is a scalar multiple of ##v2##, assuming neither vector is zero.
  • Another participant questions the definition of the inner product and suggests that proving the statement should rely on properties rather than definitions.
  • Some participants mention properties of the inner product, including conjugate symmetry, linearity, and positive-definiteness.
  • One participant relates the situation to the cosine formula, suggesting that the angle between the vectors must be either 0 or 180 degrees, indicating they are parallel.
  • Another participant proposes a method involving the vector ##u = v_2 - \frac{}{} v_1## and suggests showing that #### is zero under Cauchy-Schwarz equality.
  • A different approach is presented that involves the orthogonal projection of ##v_2## onto the subspace spanned by ##v_1##, leading to a comparison of norms and the conclusion that ##v_2 - w = 0##.
  • Several participants discuss the implications of using a Hermitian inner product in the complex case and how it affects the proof.

Areas of Agreement / Disagreement

Participants express various methods and perspectives on proving the statement, indicating that multiple approaches exist. There is no consensus on a single method or resolution of the proof, and some participants highlight the complexity of the definitions involved.

Contextual Notes

The discussion highlights the importance of understanding the properties of inner products in complex spaces, as well as the nuances in applying geometric interpretations and algebraic manipulations in proofs.

Bipolarity
Messages
773
Reaction score
2
I'm learning about Support Vector Machines and would like to recap on some basic linear algebra. More specifically, I'm trying to prove the following, which I'm pretty sure is true:
Let ##v1## and ##v2## be two vectors in an inner product space over ##\mathbb{C}##.
Suppose that ## \langle v1 , v2 \rangle = ||v1|| \cdot ||v2|| ##, i.e. the special case of Cauchy Schwarz when it is an equality. Then prove that ##v1## is a scalar multiple of ##v2##, assuming neither vector is ##0##.

I've tried using the triangle inequality and some other random stuff to no avail. I believe there's some algebra trick involved, could someone help me out? I really want to prove this and get on with my machine learning.

Thanks!

BiP
 
Physics news on Phys.org
How <v1,v2> is defined?
 
Proving this should not require the definition of the inner product, only the properties.
 
What's the difference? Which properties do you mean?
 
Conjugate symmetry, linearity in the first argument, and positive-definiteness.
 
Looks to me as another version of the cosine formula if applied to v1+v2
 
By definition, \langle v_1, v_2 \rangle = \| v_1 \| \cdot \| v_2 \| \cdot \cos(\theta) where \theta is the angle between vectors v_1 and v_2. If you also additionally know that \langle v_1, v_2 \rangle = \| v_1 \| \cdot \| v_2 \|, then the angle between the two vectors must either be 0 or 180 degrees. So they are parallel; hence one is a scalar multiple of the other.
 
That's the definition? It would be true in a real inner product space, but this one is over ℂ.
 
  • Like
Likes   Reactions: rs1n
zinq said:
That's the definition? It would be true in a real inner product space, but this one is over ℂ.

You are absolutely right! My eyes failed me, somehow.
 
  • #10
Bipolarity said:
I'm learning about Support Vector Machines and would like to recap on some basic linear algebra. More specifically, I'm trying to prove the following, which I'm pretty sure is true:
Let ##v1## and ##v2## be two vectors in an inner product space over ##\mathbb{C}##.
Suppose that ## \langle v1 , v2 \rangle = ||v1|| \cdot ||v2|| ##, i.e. the special case of Cauchy Schwarz when it is an equality. Then prove that ##v1## is a scalar multiple of ##v2##, assuming neither vector is ##0##.

I've tried using the triangle inequality and some other random stuff to no avail. I believe there's some algebra trick involved, could someone help me out? I really want to prove this and get on with my machine learning.

Thanks!

BiP

One way to do it is to consider the vector ##u = v_2 - \frac{<v1, v2>}{<v1, v1>} v_1##

Look at ##<u, u>## and show that it's zero when you have C-S equality. This also leads to a proof of the C-S inequality.
 
  • #11
To get back to the problem, though... over the complex numbers, the inner product is presumably a Hermitian inner product. So

##\begin{align*}
\| u + v \|^2 & = \langle u + v, u+v \rangle = \langle u,u \rangle + \langle u,v \rangle + \langle v,u \rangle + \langle v, v \rangle\\
& = \langle u,u \rangle + \langle u,v \rangle + \overline{\langle u,v \rangle} + \langle v, v \rangle \\
& = \langle u,u \rangle + 2 \mathrm{Re}(\langle u,v \rangle) + \langle v, v \rangle\\
& = \| u\|^2 + 2 \mathrm{Re}(\langle u,v \rangle) + \| v\|^2
\end{align*}##

Similarly,

## 0 \le \| u + \lambda v \|^2 = \| u\|^2 + 2 \mathrm{Re}(\overline{\lambda} \langle u,v \rangle) + |\lambda|^2 \| v\|^2##

Let $$\lambda = -\frac{\langle u, v\rangle }{\|v \|^2}$$ and the right hand side (above) will simplify to the C.S. inequality. Equality occurs if $$\| u + \lambda v \| = 0$$
 
  • #12
There are few possible ways of doing that. The first one is just to follow the proof of the Cauchy--Schwarz. Namely, for real ##t## consider $$\|v_1 - t v_2\|^2 = \|v_1\|^2 +t^2\|v_2\|^2 - 2t (v_1, v_2) = \|v_1\|^2 +t^2\|v_2\|^2 - 2t \|v_1\|\cdot \|v_2\| = (\|v_1\|-t\|v_2\|)^2.$$ The right hand side of this chain of equations is ##0## when ##t=\|v_1\|/\|v_2\|##. So for this ##t## you get that ##v_1-tv_2=0##, which is exactly what you need.

Another way is more geometric and probably more intuitive. You define ##w## to be the orthogonal projection of ##v_2## onto the one dimensional subspace spanned by ##v_1##, ##w= \|v_1\|^{-2} (v_2, v_1) v_1##. Then ##(v_1, v_2)= (v_1, w)## (checked by direct calculation) and ##v_2-w## is orthogonal to ##v_1## (and so to ##w##).
Therefore ##\|v_2\|^2 =\|w\|^2+\|v_2-w\|^2##.

By Cauchy--Schwarz ## (v_1, w) \le \|v_1\|\cdot \|w\|##, but on the other hand ##(v_1, w) = (v_1, v2) = \|v_1\|\cdot \|v_2\|##, so ##\|v_1\|\cdot \|v_2\| \le \|v_1\|\cdot \|w\|## and therefore ##\|v_2\|\le \|w\|##. Comparing this with ##\|v_2\|^2 =\|w\|^2+\|v_2-w\|^2## we conclude that ##v_2-w=0##.

The second proof is a bit longer, but it is more intuitive, in a sense that it is a pretty standard reasoning used when one works with orthogonal projections.
 
  • #13
Hawkeye18 said:
Another way is more geometric and probably more intuitive. You define ##w## to be the orthogonal projection of ##v_2## onto the one dimensional subspace spanned by ##v_1##, ##w= \|v_1\|^{-2} (v_2, v_1) v_1##. Then ##(v_1, v_2)= (v_1, w)## (checked by direct calculation) and ##v_2-w## is orthogonal to ##v_1## (and so to ##w##).
Therefore ##\|v_2\|^2 =\|w\|^2+\|v_2-w\|^2##.

By Cauchy--Schwarz ## (v_1, w) \le \|v_1\|\cdot \|w\|##, but on the other hand ##(v_1, w) = (v_1, v2) = \|v_1\|\cdot \|v_2\|##, so ##\|v_1\|\cdot \|v_2\| \le \|v_1\|\cdot \|w\|## and therefore ##\|v_2\|\le \|w\|##. Comparing this with ##\|v_2\|^2 =\|w\|^2+\|v_2-w\|^2## we conclude that ##v_2-w=0##.

The second proof is a bit longer, but it is more intuitive, in a sense that it is a pretty standard reasoning used when one works with orthogonal projections.

The second method is what I suggested in post #10. And, in fact, you can prove Cauchy Schwartz more intuitively this way.
 
  • #14
I see! Thank you all for your replies! I knew I had seen it somewhere, little did I know it was right there in the proof of the C-S inequality itself!

BiP
 

Similar threads

  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 8 ·
Replies
8
Views
4K
Replies
1
Views
2K
  • · Replies 5 ·
Replies
5
Views
4K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 6 ·
Replies
6
Views
4K
  • · Replies 23 ·
Replies
23
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
Replies
4
Views
2K