Uniquely Determined Linear Operator

Click For Summary
SUMMARY

The discussion centers on proving that any linear operator \(T\) on an n-dimensional linear vector space \(X\) is uniquely determined by its action on a basis \(\{x_i\}_{i=1}^{n}\) of \(X\). Participants confirm that if \(T\) maps two representations of a vector \(v\) to the same output, then the coefficients in the linear combination must be equal, establishing uniqueness. Additionally, it is concluded that \(T\) is bounded, as finite-dimensional linear transformations are inherently bounded, supported by the theorem stating that any linear transformation from a finite-dimensional space is bounded.

PREREQUISITES
  • Understanding of linear operators and vector spaces
  • Familiarity with the concepts of linear independence and spanning sets
  • Knowledge of bounded linear transformations in finite-dimensional spaces
  • Basic understanding of norms in vector spaces
NEXT STEPS
  • Study the properties of linear transformations in finite-dimensional vector spaces
  • Learn about the implications of linear independence and spanning sets in vector spaces
  • Explore the concept of boundedness in the context of linear operators
  • Investigate various norms (e.g., p-norms) and their applications in linear algebra
USEFUL FOR

Mathematicians, students of linear algebra, and anyone interested in the properties of linear transformations and vector spaces will benefit from this discussion.

Sudharaka
Gold Member
MHB
Messages
1,558
Reaction score
1
Hi everyone, :)

Here's a problem that I want to confirm my answer. Note that for the second part of the question it states, "prove that \(T\) is bonded by the above claim". I used a different method and couldn't find a method that relates the first part to prove the second.

Problem:

Suppose \(X\) is a n-dimensional linear vector space. Prove that any linear operator \(T\) on \(X\) is uniquely determined by \(\{Tx_i\}_{i=1}^{n}\) with \(\{x_i\}_{i=1}^{n}\) a basis for \(X\). Moreover, prove that \(T\) is bounded by the above claim.

My Ideas:

Suppose there are two representations of \(Tv\) where \(v\in V\). That is,

\[Tv=a_1 Tx_1+\cdots+a_n Tx_n=b_1 Tx_1+\cdots+b_n Tx_n\]

\[(a_1 Tx_1+\cdots+a_n Tx_n)-(b_1 Tx_1+\cdots+b_n Tx_n)=0\]

Since \(T\) is linear,

\[T((a_1-b_1) x_1+(a_2-b_2) x_2+\cdots+(a_n-b_n) x_n)=0\]

\[(a_1-b_1) x_1+(a_2-b_2) x_2+\cdots+(a_n-b_n) x_n=0\]

Since \(\{x_i\}_{i=1}^{n}\) is linearly independent,

\[a_i=b_i\mbox{ for all }i=1,\,2,\,\cdots,\,n\]

That is, \(T\) is uniquely determined by \(\{Tx\}_{i=1}^{n}\).

Now we shall show that \(T\) is bounded. Since \(X\) is n-dimensional, \(X\) is topologically isomorphic. That is there exist two positive constants \(c_1\) and \(c_2\) such that,

\[c_1\|x\|\leq \|Tx\|\leq c_2 \|x\|\]

for all \(x\in X\). Hence it's obvious that \(T\) is bounded.
 
Physics news on Phys.org
There is a problem with your proof:

One cannot, in general, deduce from $Tu = 0$ that $u = 0$.

I will give an example in $\Bbb R^3$, to illustrate:

Suppose $T(a,b,c) = (a,b,0)$. We have (for example), $T(1,1,1) = T(1,1,0)$, but clearly it is not the case that:

$(1,1,1) = (1,1,0)$.

In other words, $T(0,0,1) = (0,0,0)$, but $(0,0,1)$ is not the 0-vector.

What you need to show is not that $\{Tx_i\}$ is linearly independent (it may NOT be), but rather that it spans the image space. Thus we can choose some SUBSET of $\{Tx_i\}$ to be a basis for $T(X)$. We are unconcerned with elements of $X - T(X)$ since those HAVE no representation as linear combinations of the $Tx_i$.

The second half of your proof is also a bit confusing: what is it you are claiming $T(X)$ is topologically isomorphic TO? It certainly is NOT $T(X)$, unless $T$ is bijective.
 
Deveno said:
There is a problem with your proof:

One cannot, in general, deduce from $Tu = 0$ that $u = 0$.

I will give an example in $\Bbb R^3$, to illustrate:

Suppose $T(a,b,c) = (a,b,0)$. We have (for example), $T(1,1,1) = T(1,1,0)$, but clearly it is not the case that:

$(1,1,1) = (1,1,0)$.

In other words, $T(0,0,1) = (0,0,0)$, but $(0,0,1)$ is not the 0-vector.

What you need to show is not that $\{Tx_i\}$ is linearly independent (it may NOT be), but rather that it spans the image space. Thus we can choose some SUBSET of $\{Tx_i\}$ to be a basis for $T(X)$. We are unconcerned with elements of $X - T(X)$ since those HAVE no representation as linear combinations of the $Tx_i$.

Thanks so much for the reply. I understand perfectly about the fault in proof. I was rather confused by the term "uniquely determined" and thought that I should show that it's linearly independent. But then why don't they just use the word "span" instead of "uniquely determined"? Is there a difference between the two? Hope you can explain this to me. :)

Then the first part of the problem becomes quite easy, I guess. Take any vector \(v\in X\). Since \(\{x_i\}_{i=1}^{n}\) is a basis of \(X\) we have,

\[T(v)=T(a_1 x_1+\cdots+a_n x_n)=a_1 T(x_1)+\cdots+a_n T(x_n)\]

Therefore \(\{Tx_i\}_{i=1}^{n}\) spans \(\mbox{Img }T\). Am I correct?

Deveno said:
The second half of your proof is also a bit confusing: what is it you are claiming $T(X)$ is topologically isomorphic TO? It certainly is NOT $T(X)$, unless $T$ is bijective.

When I am writing the answer I took \(T\) maps \((X,\,\|\cdot\|_X)\) into \((Y,\,\|\cdot\|)_Y\). And since both are \(n-dimensional\) \(X\) and \(Y\) are topologically isomorphic. But there is a mistake in my original post. The last inequality should be,

\[c_1\|x\|_X\leq \|Tx\|_Y\leq c_2 \|x\|_X\]

I also figure out now that there's a obvious mistake here. We don't know whether \(Y\) is n-dimensional or not. I have to start from the beginning. :)

Edit: I think I got it. There's another theorem;

Let \((X,\,\|\cdot\|_X)\) and \((Y,\,\|\cdot\|)_Y\) be two normed linear spaces over \(F\). If \(X\) is finite dimensional, then any linear transformation \(T:\, X\rightarrow Y\) is bounded.

We can use this theorem to show that \(T\) is bounded. Am I correct? :)
 
I'd like to address the first part of the problem, and then we can move on to the second part.

It's not generally true for a function $f:X \to Y$ that the value of $f$ on a finite subset of $X$ completely determines $f$. For example, when $X = Y = \Bbb R$ (and these are both vector spaces), there are an infinite number of functions with:

$f(0) = a$
$f(1) = b$

no matter how we choose $a$ and $b$.

So linear functions are very special in this regard: for a vector space of dimension 2, for example, we only need the value of a linear function at 2 (linearly independent) points to know which one we have. This is essentially saying that a linear function $T: X \to X$ is completely determined by its matrix relative to any basis for $X$.

You seem to be focusing on the values of $T$ and arguing these have only "one representation". This isn't what you want to do (and it's not TRUE). What you want to do is THIS:

Suppose that $S,T$ are two linear functions with:

$S(x_i) = T(x_i),\ i = 1,2,\dots,n$.

Then, for any $v \in X$:

$S(v) = S(c_1x_1 + \cdots c_nx_n) = c_1S(x_1) + \cdots + c_nS(x_n)$

$= c_1T(x_1) + \cdots + c_nT(x_n) = T(c_1x_1 + \cdots + c_nx_n) = T(v)$

so we conclude $S = T$.

Before we move on to part 2, make sure you understand this. Also, the notion of boundedness generally requires a norm...what norm are we using for $X$ (as it is not given by the problem)?
 
Deveno said:
I'd like to address the first part of the problem, and then we can move on to the second part.

It's not generally true for a function $f:X \to Y$ that the value of $f$ on a finite subset of $X$ completely determines $f$. For example, when $X = Y = \Bbb R$ (and these are both vector spaces), there are an infinite number of functions with:

$f(0) = a$
$f(1) = b$

no matter how we choose $a$ and $b$.

So linear functions are very special in this regard: for a vector space of dimension 2, for example, we only need the value of a linear function at 2 (linearly independent) points to know which one we have. This is essentially saying that a linear function $T: X \to X$ is completely determined by its matrix relative to any basis for $X$.

You seem to be focusing on the values of $T$ and arguing these have only "one representation". This isn't what you want to do (and it's not TRUE). What you want to do is THIS:

Suppose that $S,T$ are two linear functions with:

$S(x_i) = T(x_i),\ i = 1,2,\dots,n$.

Then, for any $v \in X$:

$S(v) = S(c_1x_1 + \cdots c_nx_n) = c_1S(x_1) + \cdots + c_nS(x_n)$

$= c_1T(x_1) + \cdots + c_nT(x_n) = T(c_1x_1 + \cdots + c_nx_n) = T(v)$

so we conclude $S = T$.

Thanks for the detailed explanation. I think I am getting the intuition behind the phrase "uniquely determined" after reading your post. However there's one little point I want to clarify.

So given the set of values \(\{Tx_i\}_{i=1}^{n}\), only one linear transformation \(T\) exists with these values as it's image with respect to the basis \(\{x_i\}_{i=1}^{n}\). Is this statement correct? :)

In your proof you have taken two linear transformations \(S\) and \(T\) and shown that they are the same. However isn't it more appropriate to let, \(\{T(x_i):\,i=1,\,\cdots,\,n\}=\{S(x_j):\, j=1,\,\cdots,\,n\}\). ? I mean, for example, it could be that \(Tx_1=Sx_2,\,Tx_2=Sx_1\) and so on. Or is there a necessity that \(Tx_i=Sx_i\) for all \(i\). ? Did you get what I meant? :)

Deveno said:
Before we move on to part 2, make sure you understand this. Also, the notion of boundedness generally requires a norm...what norm are we using for $X$ (as it is not given by the problem)?

Yes, I thought about this too. That why I took a general norm without specifying anything in particular. But then after your question I tried to find a norm for this vector space and I came up with the following.

\[\|x\|=\sum_{i=1}^{n}|a_i|\]

where \(x=\sum_{i=1}^{n}a_{i}x_i\). How about that, is it correct? :)
 
Um, no...the indices have to match. Take a simple example:

Let $X = \Bbb R^2$ with:

$T = I = \text{id}_X$ (that is, $Tv = v$ for all $v \in \Bbb R^2$), and let:

$S(x,y) = (y,x)$. For the basis $B = \{(1,0),(0,1)\}$ we certainly have:

$T(1,0) = S(0,1)$
$T(0,1) = S(1,0)$, but these are clearly not the same linear transformation.

As to the norm question, there are MANY possible norms on a given vector space. It's very likely that the "usual" or "p-2 norm" (aka the Euclidean norm) is what is intended:

For $x = (x_1,\dots,x_n)$, we have:

$\displaystyle \|x\| = \left(\sum_{i = 1}^n x_i^2 \right)^{\frac{1}{2}}$

The norm you have given is also known as the "p-1 norm" or "taxi-cab norm" (or sometimes the "Manhattan norm"). We also get a norm for any real number $p \geq 1$ by taking:

$\displaystyle \|x\| = \left(\sum_{i = 1}^n x_i^p \right)^{\frac{1}{p}}$

and if we take the limit as $p \to \infty$, we get the "maximum norm":

$\|x\| = \text{max}(|x_1|,\dots,|x_n|)$.

It is not clear which of these norms is intended, of if the problem is asking you to show that $T$ is (locally) bounded for ANY norm, which is a slightly different question.
 
Deveno said:
Um, no...the indices have to match. Take a simple example:

Let $X = \Bbb R^2$ with:

$T = I = \text{id}_X$ (that is, $Tv = v$ for all $v \in \Bbb R^2$), and let:

$S(x,y) = (y,x)$. For the basis $B = \{(1,0),(0,1)\}$ we certainly have:

$T(1,0) = S(0,1)$
$T(0,1) = S(1,0)$, but these are clearly not the same linear transformation.

As to the norm question, there are MANY possible norms on a given vector space. It's very likely that the "usual" or "p-2 norm" (aka the Euclidean norm) is what is intended:

For $x = (x_1,\dots,x_n)$, we have:

$\displaystyle \|x\| = \left(\sum_{i = 1}^n x_i^2 \right)^{\frac{1}{2}}$

The norm you have given is also known as the "p-1 norm" or "taxi-cab norm" (or sometimes the "Manhattan norm"). We also get a norm for any real number $p \geq 1$ by taking:

$\displaystyle \|x\| = \left(\sum_{i = 1}^n x_i^p \right)^{\frac{1}{p}}$

and if we take the limit as $p \to \infty$, we get the "maximum norm":

$\|x\| = \text{max}(|x_1|,\dots,|x_n|)$.

It is not clear which of these norms is intended, of if the problem is asking you to show that $T$ is (locally) bounded for ANY norm, which is a slightly different question.

Thanks so much for all your help. :) I think I understand this completely now. I think the question expects us to show that the linear map is bounded for any norm which is in fact a theorem given in our recommended textbook. :)
 
For FINITE-DIMENSIONAL vector spaces, (local) boundedness is a consequence of continuity (at 0, and thus everywhere). I remark in passing that the definition of continuity is also dependent upon the norm (one uses the norm to establish a metric, and then uses the definition of continuity in a metric space...otherwise you lack quantities to be less than your epsilons and deltas).

In particular, by choosing a delta-ball centered at the origin that is small enough radius, we can ensure that $\|T(x)\| \leq 1$ for all $x$ with $\|x\| \leq \delta$ (in other words, we pick an epsilon equal to 1).

Then for ANY $v \in X$, we have:

$\displaystyle \|T(v)\| = \left\|\frac{\|v\|}{\delta}T\left(\delta \frac{v}{\|v\|}\right)\right\|$

$\displaystyle = \frac{\|v\|}{\delta}\left\|T\left(\delta \frac{v}{\|v\|}\right)\right\|$

$\displaystyle \leq \frac{1}{\delta}\|v\|$

so we may take $M = \dfrac{1}{\delta}$, which established local boundedness, and depends only on the generic property every norm has, namely:

$\|\alpha v\| = |\alpha|\cdot \|v\|$
 

Similar threads

  • · Replies 27 ·
Replies
27
Views
2K
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 0 ·
Replies
0
Views
932
  • · Replies 25 ·
Replies
25
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 23 ·
Replies
23
Views
2K
  • · Replies 4 ·
Replies
4
Views
2K