What are the properties of affine bases in $\mathbb{R}^n$?

  • Context: MHB 
  • Thread starter Thread starter mathmari
  • Start date Start date
  • Tags Tags
    Bases
Click For Summary

Discussion Overview

The discussion revolves around the properties of affine bases in $\mathbb{R}^n$, focusing on geometric descriptions, transformations, and the implications of isometries. Participants explore various aspects of affine bases, including their definitions, transformations under linear mappings, and conditions for isometries.

Discussion Character

  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • Some participants describe an affine basis in $\mathbb{R}^n$ as a set of points $(p_0, p_1, \ldots, p_n)$ such that the vectors $p_1 - p_0, p_2 - p_0, \ldots, p_n - p_0$ form a basis of $\mathbb{R}^n$.
  • It is proposed that for any vector $v \in \mathbb{R}^n$, the set $(p_0 + v, p_1 + v, \ldots, p_n + v)$ also constitutes an affine basis of $\mathbb{R}^n$.
  • Participants discuss the effect of an invertible matrix $a$ on the affine basis, suggesting that $(ap_0, ap_1, \ldots, ap_n)$ remains an affine basis due to the linear independence of the transformed vectors.
  • There is a suggestion that isometries can be expressed as a combination of orthogonal transformations and translations, which may help in proving that $(\beta(p_0), \ldots, \beta(p_n))$ is an affine basis.
  • Some participants express uncertainty about how to show that if $\beta(p_i) = p_i$ for all $i$, then $\beta$ must be the identity transformation.
  • There is a discussion about the implications of the condition $p_0 = 0$ in proving that $a$ must be the identity matrix when $p_i$ form a basis of $\mathbb{R}^n$.
  • Participants explore the relationship between the affine basis and the properties of the transformation $\beta$, questioning how to derive that $v = 0$ without additional conditions.

Areas of Agreement / Disagreement

Participants generally agree on the definitions and properties of affine bases, but there is ongoing debate regarding the implications of isometries and the conditions under which certain transformations hold. The discussion remains unresolved on some aspects, particularly concerning the necessity of certain conditions for proving claims about isometries.

Contextual Notes

Some participants note that the arguments depend on the assumption that the $p_i$ form a basis of $\mathbb{R}^n$, which is crucial for establishing the linear independence of the vectors involved. There are also unresolved questions regarding the implications of transformations and the specific conditions required for certain results.

Who May Find This Useful

This discussion may be useful for students and researchers interested in linear algebra, geometry, and the properties of transformations in vector spaces, particularly those exploring the concepts of affine spaces and isometries.

mathmari
Gold Member
MHB
Messages
4,984
Reaction score
7
Hey! :giggle:

Let $1\leq n\in \mathbb{N}$ and $(p_0,\ldots , p_n)$ an affine basis (that means that the vectors $p_1-p_0, p_2-p_0,\ldots ,p_n-p_0$ build a basis of $\mathbb{R}^n$.
(a) Give a geometric description of affine bases of $\mathbb{R}^n$ for $1\leq n\leq 3$.
(b) For all $v\in \mathbb{R}^n$ show that $(p_0+v,\ldots , p_n+v)$ is an affine basis of $\mathbb{R}^n$.
(c) Let $a$ be an invertible matrix. Then show that $(ap_0,\ldots , ap_n)$ is an affine basis of $\mathbb{R}^n$.
(d) For each isometry $\beta\in \text{Isom}(\mathbb{R}^n)$ show that $(\beta (p_0),\ldots , \beta( p_n))$ an affine basis of $\mathbb{R}^n$.
(e) Let $p_0=0$ and $\beta\in \text{Isom}(\mathbb{R}^n)$, with $\beta (p_i)=p_i$ for all $0\leq i\leq n$. Then show that $\beta=\text{id}_{\mathbb{R}^n}$.
(f) Let $\beta\in \text{Isom}(\mathbb{R}^n)$, with $\beta (p_i)=p_i$ for all $0\leq i\leq n$. Then show that $\beta=\text{id}_{\mathbb{R}^n}$.For (b) I have done the following :
We have that the vectors \begin{align*}&(p_1+v)-(p_0-v)=p_1+v-p_0-v=p_1-p_0 \\ &(p_2+v)-(p_0-v)=p_2+v-p_0-v=p_2-p_0 \\ &\ldots \\ &(p_n+v)-(p_0-v)=p_n+v-p_0-v=p_n-p_0 \end{align*}
build a basis of $\mathbb{R}^n$, since $(p_0, p_1, \ldots ,p_n)$ is an affine basis of $\mathbb{R}^n$.

Therefore $(p_0+v, p_1+v, \ldots ,p_n+v)$ is an affine basis of $\mathbb{R}^n$. For (c) I have done the following :
We have that \begin{align*}&ap_1-ap_0=a(p_1-p_0) \\ &ap_2-ap_0=a(p_2-p_0) \\ &\ldots \\ &ap_n-ap_0=a(p_n-p_0) \end{align*}
We have that \begin{align*}\lambda_1[a(p_1-p_0)]+\lambda_2[a(p_2-p_0)]+\cdots +\lambda_n[a(p_n-p_0)]=0 &\Rightarrow a\left (\lambda_1(p_1-p_0)+\lambda_2(p_2-p_0)+\cdots +\lambda_n(p_n-p_0)\right )=0 \ \\ & \overset{a\in \text{GL}_n(\mathbb{R})}{\Rightarrow } \ a^{-1}a\left (\lambda_1(p_1-p_0)+\lambda_2(p_2-p_0)+\cdots +\lambda_n(p_n-p_0)\right )=a^{-1}\cdot 0 \\ & \Rightarrow \lambda_1(p_1-p_0)+\lambda_2(p_2-p_0)+\cdots +\lambda_n(p_n-p_0)=0\end{align*}
Since $(p_0, p_1, \ldots ,p_n)$ is an affine basis of $\mathbb{R}^n$ the $p_1-p_0, \ p_2-p_0, \ \ldots , \ p_n-p_0$ build a basis of $\mathbb{R}^n$ and so they are linearly independent, and so $\lambda_1=\lambda_2=\ldots =\lambda_n=0$.

So the $n$ linearly independent vectors $a(p_1-p_0), \ a(p_2-p_0), \ \ldots , \ a(p_n-p_0)$ in the $n$-dimensional space are a basis. So $(p_0+v, p_1+v, \ldots ,p_n+v)$ is anaffine basis of $\mathbb{R}^n$.Is that correct so far? :unsure:

Could you give me a hint for he remaining parts? For example for the isometries the distances are preserved but canwe say that the differences of $\beta(p_i)-\beta(p_0)$ are equal to $p_i-p_0$ ? :unsure:
 
Physics news on Phys.org
mathmari said:
Could you give me a hint for he remaining parts? For example for the isometries the distances are preserved but canwe say that the differences of $\beta(p_i)-\beta(p_0)$ are equal to $p_i-p_0$ ?
Hey mathmari!

Every isometry in $\mathbb R^n$ with the usual metric can be written as an orthogonal transformation plus a translation.
We can use that for (d), (e), and (f), can't we? 🤔
 
Klaas van Aarsen said:
Every isometry in $\mathbb R^n$ with the usual metric can be written as an orthogonal transformation plus a translation.
We can use that for (d), (e), and (f), can't we? 🤔

How do we use that? I got stuck right now. :unsure:
 
mathmari said:
How do we use that? I got stuck right now.
Use that $\beta(x)=ax+v$ where $a$ is an invertible matrix and use the same logic as in (b) and (c)? 🤔
We also have that $a^Ta=\text{id}_{\mathbb R^n}$.
 
Klaas van Aarsen said:
Use that $\beta(x)=ax+v$ where $a$ is an invertible matrix and use the same logic as in (b) and (c)? 🤔
We also have that $a^Ta=\text{id}_{\mathbb R^n}$.

For (d) I have done the following :

We have that \begin{align*}&\beta (p_1)-\beta (p_0)=(ap_1+v)-(ap_0+v)=ap_1-ap_0=a(p_1-p_0) \\ &\beta (p_2)-\beta (p_0)=(ap_2+v)-(ap_0+v)=ap_2-ap_0=a(p_2-p_0) \\ &\ldots \\ &\beta (p_n)-\beta (p_0)=(ap_n+v)-(ap_0+v)=ap_n-ap_0=a(p_n-p_0) \end{align*}
We have that \begin{align*}&\lambda_1[a(p_1-p_0)]+\lambda_2[a(p_2-p_0)]+\cdots +\lambda_n[a(p_n-p_0)]=0 \\ &\Rightarrow a\left (\lambda_1(p_1-p_0)+\lambda_2(p_2-p_0)+\cdots +\lambda_n(p_n-p_0)\right )=0 \ \\& \overset{a\in \text{GL}_n(\mathbb{R})}{\Rightarrow } \ a^{-1}a\left (\lambda_1(p_1-p_0)+\lambda_2(p_2-p_0)+\cdots +\lambda_n(p_n-p_0)\right )=a^{-1}\cdot 0 \\ & \Rightarrow \lambda_1(p_1-p_0)+\lambda_2(p_2-p_0)+\cdots +\lambda_n(p_n-p_0)=0\end{align*}
Since $(p_0, p_1, \ldots ,p_n)$ is an affine basis $\mathbb{R}^n$, it follows that $p_1-p_0, \ p_2-p_0, \ \ldots , \ p_n-p_0$ is a basis of $\mathbb{R}^n$, and so these are linearly independent, so $\lambda_1=\lambda_2=\ldots =\lambda_n=0$.
So it follows that the $n$ linearly independent vectors $a(p_1-p_0), \ a(p_2-p_0), \ \ldots , \ a(p_n-p_0)$, i.e. $\beta (p_1)-\beta (p_0), \ \beta (p_2)-\beta (p_0), \ \ldots , \ \beta (p_n)-\beta (p_0)$, in $n$-dimensional space form a basis.
Therefore $(\beta (p_0), \beta (p_1), \ldots ,\beta (p_n))$ is an affine basis of $\mathbb{R}^n$.
For (e) I have done the following :

We have that $\beta(p_i)=p_i \Rightarrow ap_i+v=p_i\Rightarrow (a-u)p_i=-v$.
How do we continue? :unsure:
 
mathmari said:
We have that $\beta(p_i)=p_i \Rightarrow ap_i+v=p_i\Rightarrow (a-u)p_i=-v$.
How do we continue?
Suppose we substitute $p_0=0$? :unsure:
 
Klaas van Aarsen said:
Suppose we substitute $p_0=0$? :unsure:

From $\beta(p_i)=p_i \Rightarrow ap_i+v=p_i\Rightarrow (a-u)p_i=-v$ for $i=0$ we have $p_0=0$ so we get $(a-u)\cdot 0=-v \Rightarrow v=0$. Therefore $ap_i=p_i$. Do we have to show that $a$ is the identity matrix? :unsure:
 
mathmari said:
From $\beta(p_i)=p_i \Rightarrow ap_i+v=p_i\Rightarrow (a-u)p_i=-v$ for $i=0$ we have $p_0=0$ so we get $(a-u)\cdot 0=-v \Rightarrow v=0$. Therefore $ap_i=p_i$. Do we have to show that $a$ is the identity matrix?
Yep. (Nod)
 
Klaas van Aarsen said:
Yep. (Nod)

Do we consider for that the system $(a-I)p_i=0$ ? Since the $p_i$'s cannot be zero because the form an affine basis, it follows that $a-I=0$, i.e. $a=I$.

Is that correct? :unsure:
 
  • #10
mathmari said:
Do we consider for that the system $(a-I)p_i=0$ ? Since the $p_i$'s cannot be zero because the form an affine basis, it follows that $a-I=0$, i.e. $a=I$.

Is that correct?
It's not sufficient that the $p_i$ are non-zero. We actually need that the $p_i$ for $n=1,\ldots,n$ form a basis of $\mathbb R^n$.
It follows from there that $a$ must indeed be the identity matrix. 🤔
 
  • #11
Klaas van Aarsen said:
It's not sufficient that the $p_i$ are non-zero. We actually need that the $p_i$ for $n=1,\ldots,n$ form a basis of $\mathbb R^n$.
It follows from there that $a$ must indeed be the identity matrix. 🤔

We have that the $p_i$ for $n=1,\ldots,n$ form a basis of $\mathbb R^n$. But how does it follow that $a$ must indeed be the identity matrix? :unsure:
 
  • #12
mathmari said:
We have that the $p_i$ for $n=1,\ldots,n$ form a basis of $\mathbb R^n$. But how does it follow that $a$ must indeed be the identity matrix? :unsure:
Suppose we put the $p_i$ as column vectors in a basis matrix $b$, which is invertible.
Then $ap_i=p_i\implies ab=b \implies abb^{-1}=bb^{-1}\implies a=I$. 🤔
 
  • #13
Klaas van Aarsen said:
Suppose we put the $p_i$ as column vectors in a basis matrix $b$, which is invertible.
Then $ap_i=p_i\implies ab=b \implies abb^{-1}=bb^{-1}\implies a=I$. 🤔

Ahh I see!

So at (f) we do that in a similar way:
We have that $\beta(p_i)=p_i \Rightarrow ap_i+v=p_i\Rightarrow (a-u)p_i=-v$. But how do we get that $v=0$ ? We don't have any additional condition that we could use as before that $p_0=0$.
Could you give me a hint? :unsure:
 
  • #14
What is $\beta(p_i-p_0)$? 🤔
 
  • #15
Klaas van Aarsen said:
What is $\beta(p_i-p_0)$? 🤔

We have that $\beta (p_i-p_0)=a(p_i-p_0)+v=ap_i-ap_0+v$, or not? :unsure:
 
  • #16
How about $\beta p_i - \beta p_0$?
And in particular for $i=0$? 🤔:
 

Similar threads

  • · Replies 28 ·
Replies
28
Views
3K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 23 ·
Replies
23
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 10 ·
Replies
10
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 12 ·
Replies
12
Views
2K
  • · Replies 15 ·
Replies
15
Views
2K