Inner product - Analysis in Rn problem

In summary: We can see this in the following way:Given a vector ##x_1## we can write it as the sum of two vectors, one being the projection of ##x_1## onto the vector ##x## and the other being the projection of the same vector ##x_1## onto a vector orthogonal to ##x##. In other words we can write ##x_1 = c_1x + c_2(x_1 - c_1x)##.
  • #1
Onezimo Cardoso
11
2

Homework Statement


Let ##x,y \in \mathbb{R^n}## not null vectors. If for all ##z \in \mathbb{R^n}## that is orthogonal to ##x## we have that ##z## is also orthogonal to ##y##, prove that ##x## and ##y## are multiple of each other.

Homework Equations


We can use that fact that ##<x , y-\frac{<x,y>}{|x|^2}x> = 0##

The Attempt at a Solution


If ##x=y## the result follow directly. So let us suppose that ##x\neq y##.
Using the fact that ##<x , y-\frac{<x,y>}{|x|^2}x> = 0## and the hypothesis we can affirm that ##<y , y-\frac{<x,y>}{|x|^2}x> = 0##.
Further, by the fact that ##<x , y-\frac{<x,y>}{|x|^2}x> = 0## and ##<y , y-\frac{<x,y>}{|x|^2}x> = 0## we have:
$$<x , y-\frac{<x,y>}{|x|^2}x> = <y , y-\frac{<x,y>}{|x|^2}x> $$
$$\Rightarrow <x , y-\frac{<x,y>}{|x|^2}x> - <y , y-\frac{<x,y>}{|x|^2}x> = 0$$
$$\Rightarrow <x - y , y-\frac{<x,y>}{|x|^2}x> =0 \quad (*)$$

Until here everything seems great. But now...

Once ##x\neq y## , second the source I got this solution, by ##(*)## we can affirm that:
$$y-\frac{<x,y>}{|x|^2}x = 0$$
$$\Rightarrow y = \alpha x$$
Where ##\alpha = \frac{<x,y>}{|x|^2}x ##

But the fact is if we have ##<x,y>=0## not necessarily we must have ##x=0## or ##y=0##.
Can anyone see if I could not see a relevant fact in the argument or someone have another look for this problem?

Thanks in advance.
 
Last edited:
Physics news on Phys.org
  • #2
Onezimo Cardoso said:

Homework Statement


Let ##x,y \in \mathbb{R^n}## not null vectors. If for all ##z \in \mathbb{R^n}## that is orthogonal to ##x## we have that ##z## is also orthogonal to ##y##, prove that ##x## and ##y## are multiple of each other.

Homework Equations


We can use that fact that ##<x , y-\frac{<x,y>}{|x|^2}x> = 0##

The Attempt at a Solution


If ##x=y## the result follow directly. So let us suppose that ##x\neq y##.
Using the fact that ##<x , y-\frac{<x,y>}{|x|^2}x> = 0## and the hypothesis we can affirm that ##<y , y-\frac{<x,y>}{|x|^2}x> = 0##.
Further, by the fact that ##<x , y-\frac{<x,y>}{|x|^2}x> = 0## and ##<y , y-\frac{<x,y>}{|x|^2}x> = 0## we have:
$$<x , y-\frac{<x,y>}{|x|^2}x> = <y , y-\frac{<x,y>}{|x|^2}x> $$
$$\Rightarrow <x , y-\frac{<x,y>}{|x|^2}x> - <y , y-\frac{<x,y>}{|x|^2}x> = 0$$
$$\Rightarrow <x - y , y-\frac{<x,y>}{|x|^2}x> =0 \quad (*)$$

Until here everything seems great. But now...

Once ##x\neq y## , second the source I got this solution, by ##(*)## we can affirm that:
$$y-\frac{<x,y>}{|x|^2}x = 0$$
$$\Rightarrow y = \alpha x$$
Where ##\alpha = \frac{<x,y>}{|x|^2}x ##

But the fact is if we have ##<x,y>=0## not necessarily we must have ##x=0## or ##y=0##.
Can anyone see if I could not see a relevant fact in the argument or someone have another look for this problem?

Thanks in advance.
Why can you assume
Using the fact that ##<x , y-\frac{<x,y>}{|x|^2}x> = 0##
##<x , y-\frac{<x,y>}{|x|^2}x>= <x,y>-<x,\alpha x>=<x,y>=0\,,## so you start with orthogonal vectors and end up with orthogonal vectors. Nothing happened in between.

Which results about basis and / or orthogonal basis do you know? E.g. you could write ##\mathbb{R}^n=\mathbb{R}x\oplus \mathbb{R}^{n-1}=\mathbb{R}y\oplus \mathbb{R}^{n-1}## with orthogonal subspaces. Your condition is:
$$
\forall \, z\in \mathbb{R}^n\, \text{ with } \,\langle x , z\rangle = 0 \, : \, (\,\langle y,z \rangle = 0 \Longrightarrow x \sim y\,)
$$
The fact, that it holds for all such ##z## has somehow to be used. It cannot work with a single one.
 
  • #3
fresh_42 said:
Why can you assume
##<x , y-\frac{<x,y>}{|x|^2}x>= <x,y>-<x,\alpha x>=<x,y>=0\,,## so you start with orthogonal vectors and end up with orthogonal vectors. Nothing happened in between.

Which results about basis and / or orthogonal basis do you know? E.g. you could write ##\mathbb{R}^n=\mathbb{R}x\oplus \mathbb{R}^{n-1}=\mathbb{R}y\oplus \mathbb{R}^{n-1}## with orthogonal subspaces. Your condition is:
$$
\forall \, z\in \mathbb{R}^n\, \text{ with } \,\langle x , z\rangle = 0 \, : \, (\,\langle y,z \rangle = 0 \Longrightarrow x \sim y\,)
$$
The fact, that it holds for all such ##z## has somehow to be used. It cannot work with a single one.

Ok I got your point.
But let me repharse the proof above and try to express myself better in order to show that the only problem with the proof above (at least for me) is the fact that ##(*)## and ##x-y \neq 0## implies that ##y-\frac{<x,y>}{|x|^2}x = 0##.

Let's begin:

Let us suppose that $$\forall z \in \mathbb{R}^n ; <x,z>=0 \Rightarrow <y,z>=0$$
Particularly, if we consider ##z_0=y-\frac{<x,y>}{|x|^2}x## , we have direclty that ##<x,z_0> = 0##. So by the hypothesis written just above we must have that ##<y,z_0>=0##, i.e. ##<y,y-\frac{<x,y>}{|x|^2}x>=0##... (And this proof goes on until stuck in ##(*)##)...

...
Notice that geometrically this really makes sense:
upload_2018-7-10_12-1-8.png
 

Attachments

  • upload_2018-7-10_12-1-8.png
    upload_2018-7-10_12-1-8.png
    5.8 KB · Views: 385
Last edited:
  • #4
Onezimo Cardoso said:
Ok I got your point.
But let me repharse the proof above and try to express myself better in order to show that the only problem with the proof above (at least for me) is the fact that ##(*)## and ##x-y \neq 0## implies that ##y-\frac{<x,y>}{|x|^2}x = 0##.

Let's begin:

Let us suppose that $$\forall z \in \mathbb{R}^n ; <x,z>=0 \Rightarrow <y,z>=0$$
This is not what you have. From ##\forall z \in \mathbb{R}^n ; <x,z>=0## it follows immediately that ##x=0## because the inner product is not degenerated and thus ##\langle x,z \rangle = \langle 0,z\rangle = 0## is trivial and always true. And because it is a tautology, it will not help you.
Particularly, if we consider ##z_0=y-\frac{<x,y>}{|x|^2}x## , we have direclty that ##<x,z_0> = 0##. So by the hypothesis written just above we must have that ##<y,z_0>=0##, i.e. ##<y,y-\frac{<x,y>}{|x|^2}x>=0##... (And this proof goes on until stuck in ##(*)##)...
Why can you assume ##\langle x,z_0\rangle=\langle x,y\rangle =0\,?## If you start with orthogonal ##x,y## they barely can be linear dependent.

I'm afraid you will have to use an entire basis, preferably orthogonal, but I don't know whether you can already assume it to exist. I would have placed your exercise behind the Gram-Schmidt algorithm in a book, however, I don't know your book and therefore not what you already can assume as given. That's what our section 2 of the template is meant for!
 
  • Like
Likes Onezimo Cardoso
  • #5
upload_2018-7-10_17-13-2.png
 

Attachments

  • upload_2018-7-10_17-13-2.png
    upload_2018-7-10_17-13-2.png
    6.6 KB · Views: 756
  • Like
Likes Onezimo Cardoso
  • #6

Attachments

  • upload_2018-7-10_12-35-30.png
    upload_2018-7-10_12-35-30.png
    16.3 KB · Views: 321
  • #7
The condition should read: Let ##\mathfrak{x},\mathfrak{y} \in \mathbb{R}^n -\{0\}##.
\begin{align*}
&\text{IF } &\mathfrak{y} \perp U:=\{\,\mathfrak{z}\,|\,\langle \mathfrak{x},\mathfrak{z}\rangle=0\,\} = (\mathbb{R}\mathfrak{x})^\perp \cong \mathbb{R}^{n-1} \\[6pt]
&\text{THAT IS } &\langle \mathfrak{y},\mathfrak{z}\rangle =0 \,\,\forall \, \mathfrak{z}\in U \subsetneq \mathbb{R}^n \\[6pt]
&\text{THEN } & \mathfrak{y} \sim \mathfrak{x} \\[6pt]
&\text{THAT IS } & \exists \,\alpha \in \mathbb{R}-\{0\} \, : \, \mathfrak{y}=\alpha \cdot \mathfrak{x}
\end{align*}
or for short, you have to show ##(\operatorname{span}\{\,x\,\}^\perp)^\perp = \operatorname{span}\{\,x\,\}\,.##

I hope I haven't confused the letters, but it was you who started with fraktur. Anyway, what you can see is, that a single ##z## is irrelevant, and you have to use any resp. all from the subspace ##U## of all vectors perpendicular to ##x##, which is not the entire vector space, but a subspace of codimension one. This was the reason I asked for the propositions you already know. Otherwise you'll have to repeat the Gram-Schmidt procedure in a way, or use induction on ##n## which is always a good idea when dealing with finite dimensional vector spaces. But in any case, the eintire ##U## is needed and the fact that it has codimension one, for otherwise we will always find a direction which we haven't used yet, and which is perpendicular to all of them; as shown in my modification of your drawing.
 
  • Like
Likes Onezimo Cardoso
  • #8
I think Fresh may have been alluding to this, but I'll add that there's a very nice way to do this using Projectors, in particular a well chosen real symmetric ##\mathbf P## that obeys ##\mathbf P^2 = \mathbf P## would make this a 2-4 line exercise.

On the other hand, if OP has never heard terms like idempotent or Projector, then ignore this comment.
 
  • #9
do you have any dimension theory? specifically do you know that the orthogonal complement of an r dimensional subspace has dimension n-r? if so then both vectors x and y belong to the same one dimensional subspace under your hypothesis. i.e. (xperp)perp = y(perp)perp is one dimensional and contains both x and y.
 
  • Like
Likes Onezimo Cardoso
  • #10
mathwonk said:
do you have any dimension theory? specifically do you know that the orthogonal complement of an r dimensional subspace has dimension n-r? if so then both vectors x and y belong to the same one dimensional subspace under your hypothesis. i.e. (xperp)perp = y(perp)perp is one dimensional and contains both x and y.

Great mathwonk! I'm just arranging a complete solution using your argument in order to help others that could reach this question.

Let ##x,y \in \mathbb{R}^n## not null vectors.
Let us suppose that for all ##z \in \mathbb{R}^n## which is orthogonal to ##x## it is true that ##z## is ortogonal to ##y##.
Let us considerer the orthogonal subspaces of ##x## and ##y## respectively:
$$(x^{\perp})= \{z \in \mathbb{R}^n ; <z,x>=0\}$$
$$(y^{\perp})=\{z \in \mathbb{R}^n ; <z,y>=0\}$$
From our hypothesis we have that ##(x^{\perp}) \subset (y^{\perp})##. But ##dim(x^{\perp})=dim(y^{\perp})##, then ##(x^{\perp}) = (y^{\perp})##.
Let ##S=[x,y]## be the span of ##\{ x,y \}##. Let ##(S^{\perp})## be the orthogonal complement of ##S##.
From our hypothesis we have ##(x^{\perp}) = (y^{\perp})=(S^{\perp})##.
Then ##dim((S^{\perp}))=n-1##.
But
$$dim(S)+dim((S^{\perp}))=n$$
$$\Rightarrow dim(S)=1$$
Therefore, ##x## and ##y## are linear dependent.
 
  • Like
Likes Delta2

1) What is an inner product in Rn?

An inner product in Rn is a mathematical operation that takes two vectors in n-dimensional space and produces a scalar value. It is used to measure the geometric properties of vectors, such as their length and angle between them.

2) How is an inner product calculated in Rn?

In Rn, the inner product is calculated by taking the dot product of two vectors, which involves multiplying the corresponding elements of the vectors and then summing them up. This can be expressed as a formula: a · b = a1b1 + a2b2 + ... + anbn, where a and b are vectors in Rn.

3) What is the significance of the inner product in Rn?

The inner product in Rn has many applications in mathematics and physics, such as in vector calculus, linear algebra, and quantum mechanics. It allows us to measure the similarity or orthogonality of vectors, and it also provides a way to define the length and angle between vectors in n-dimensional space.

4) Can the inner product be negative in Rn?

Yes, the inner product in Rn can be negative. This happens when the angle between two vectors is greater than 90 degrees, indicating that the vectors are pointing in opposite directions. In this case, the dot product will result in a negative value.

5) What is the difference between an inner product and a norm in Rn?

An inner product and a norm are both mathematical operations that involve vectors in Rn. However, an inner product produces a scalar value, while a norm produces a non-negative value (usually representing the length of a vector). Additionally, an inner product is a more general concept, while a norm is a specific type of inner product.

Similar threads

  • Calculus and Beyond Homework Help
Replies
5
Views
626
  • Calculus and Beyond Homework Help
Replies
5
Views
767
  • Calculus and Beyond Homework Help
Replies
10
Views
387
  • Calculus and Beyond Homework Help
Replies
5
Views
301
  • Calculus and Beyond Homework Help
Replies
10
Views
831
  • Calculus and Beyond Homework Help
Replies
18
Views
1K
  • Calculus and Beyond Homework Help
Replies
7
Views
695
  • Calculus and Beyond Homework Help
Replies
9
Views
1K
  • Calculus and Beyond Homework Help
Replies
3
Views
525
  • Calculus and Beyond Homework Help
Replies
1
Views
463
Back
Top