Struggling with change of basis

In summary, the conversation discusses the use of change of basis matrices and coordinate systems for linear transformations. The example of a simple reflection through a plane is used to show how changing the basis can make it easier to find the transformation matrix. The concept of eigenvectors is also introduced, which leads to the use of a diagonal matrix in the Jordan normal form as a canonical way to express the transformation. The conversation also touches on the importance of eigenvectors in applications of linear algebra and the fact that the choice of basis does not affect the equations for the transformation. However, the listener still has a blind spot in understanding how the vectors were "jumped" from one basis to another without using eigenvectors.
  • #1
ognik
643
2
Hi, sadly my textbook assumes a knowledge I didn't have, of change of basis matrices & coordinate systems for linear transformations; so I have been trolling around the web to fill in the gaps as best I can.

I have an open post that has no replies - http://mathhelpboards.com/linear-abstract-algebra-14/follow-basis-question-16258.html - I would really appreciate if someone could at least check the 5 hesitant assertions I make in that post, even a yes/no would help please.

----------------------------------

I've moved on as best I can anyway, been working through an example from Khan Academy - but there is one aspect I just don't get.

The example transformation is a simple reflection through a plane V, the example sets out to show that changing the basis can make it easier to find the transformation matrix A. So we have $T(\vec{x})=A\vec{x}$, and an orthonormal basis in $R^3 , B=\left\{ v_1, v_2, v_3 \right\}$. If we allow $v_1, v_2 $ to span the plane V, then the transformation for these basis vectors would be $ T(\vec{v_1})= \vec{v_1}, T(\vec{v_2})= \vec{v_2}, T(\vec{v_3})= -\vec{v_3} $.

The example suggests that a change of basis matrix C comprised of the vectors of B is a 'natural' basis to use - 1st questions, is it always likely to be the best basis to use? We could use any orthonormal basis in $R^3$? (In this case it happens to simplify the transformation, what if it didn't?)

So I follow we want to find $T(\vec{x})=CDC^{-1} \vec{x}$ and from there, $A=CDC^T $

The example then says $ \vec{v_1}=1.\vec{v_1} + 0.\vec{v_2} +0.\vec{v_3}$ (ok), and then $ \left[ \vec{v_1} \right]_B = \begin{bmatrix}1\\0\\ 0 \end{bmatrix} etc.$

I'm just having a blank spot with this for some reason and maybe a different perspective on it would help me 'click'. I can see they have written $\vec{v}$ as a linear combination of the vectors in B, so the coordinates must be w.r.t. B.
But somehow we seem to have jumped from the vectors being in the standard basis (see $ T(\vec{v_1})= \vec{v_1}, T(\vec{v_2})= \vec{v_2}, T(\vec{v_3})= -\vec{v_3} $ above) - to those same vectors also being in B; for whatever reason, this just stops my mind at this point. (If I accept it, the rest of the example is easy to follow). Try to convince me please :-)
 
Physics news on Phys.org
  • #2
As far as the vector space is concerned, ALL bases are equally good.

But we humans find some calculations more difficult than others: for example $37 + 0$ is easier to evaluate than $37 + 41$ (I can answer the first in a fraction (say half) of a second, the latter takes me 2-3 seconds, a sixfold decrease in efficiency).

Now matrices, being the beasts they are, can require a LOT of calculations to compute a matrix product (sums are not so bad). It would be nice if multiplying matrices were as easy as multiplying numbers (scalars).

One way this can happen, is if matrices are DIAGONAL. Now if it were to happen that the matrix of a linear transformation was diagonal, calculation with that matrix (in that basis) would be a LOT easier. It turns out that this isn't always possible, and to determine when it IS possible, we need to study things called eigenvectors.

So, although the Khan Academy video isn't really explicitly saying as much, what you have in the reflection example is a case where the basis vectors are chosen to be eigenvectors.

Indeed, vectors in the plane of reflection are all eigenvectors with a eigenvalue of $1$, this is to say they are left unchanged by the reflection. The $v_3$ is then a basis for what is called the orthogonal complement of the plane, which can be shown to be "independent" of the plane in the sense that:

$\Bbb R^3 = P \oplus \langle v_3\rangle$.

I do not know if your text has covered direct sums or complementary subspaces, yet.

In any case, if $B = \{v_1,v_2,v_3\}$, then:

$[T]_B = \begin{bmatrix}1&0&0\\0&1&0\\0&0&-1\end{bmatrix}$

which is indeed a diagonal matrix, and easy for us simple humans to do calculations with.

Is this the "best" way to represent $T$? That's not really the right question-it's a CANONICAL way to express $T$, in what is known as JORDAN NORMAL FORM. $B$ forms an eigenbasis for $\Bbb R^3$ with respect to $T$, which turns out to be important for many APPLICATIONS of linear algebra (where calculation difficulty matters more).

One small point: if $Tv = \lambda v$ for some scalar $\lambda$, this is true no matter WHICH basis we express $T$ and $v$ in:

Standard basis- using the matrix $A$ for $T$:

$Tv = Av = \lambda v$.

Another basis - using the matrix $P^{-1}AP$ for $T$:

$[Tv]_B = P^{1}AP[v]_B = P^{-1}A(v) = P^{-1}(\lambda v) = \lambda P^{-1}v = \lambda[v]_B = [\lambda v]_B$

so the apparent "switching back and forth" between bases in the videos doesn't affect the equations:

$T(v_1) = v_1,\ T(v_2) = v_2$, and $T(v_3) = -v_3$.
 
  • #3
Yes indeed, the next section from Khan was eigenvectors, so you correctly anticipated that. I've worked through eigenvectors now, but it doesn't help with the blind spot I mentioned - I can find the diagonal matrix using eigenvalues, but I just can't see/accept how it was done without eifenvalues, ie: $ \vec{v_1}=1.\vec{v_1} + 0.\vec{v_2} +0.\vec{v_3}$ (ok), and then $ \left[ \vec{v_1} \right]_B = \begin{bmatrix}1\\0\\ 0 \end{bmatrix} etc.$

It seems to me this somehow jumps those vectors from one basis to another, using itself...
 
  • #4
In the basis $B = \{v_1,v_2,v_3\}$ we have $[v_1]_B = (1,0,0)$ (these are the $B$-coordinates of $v_1$, which in another post I said I like to denote as $[1,0,0]_B$ so as to avoid confusion with $e_1$).

The argument that $[T]_B([v_1]_B) = [Tv_1]_B = [v_1]_B$ is based on GEOMETRY, reflecting 3-space through a plane leaves the plane itself unchanged. In the basis $B$, therefore, the first column of the matrix for $T$ is $(1,0,0)^t$.

Just to be clear- a triple of numbers (which we associate naturally with a POINT in $\Bbb R^3$) is just a triple of numbers:

$(3,2,1)$ might mean: 3 feet east, 2 feet north, and 1 foot up, or it might mean: 3 miles east, 2 yards north and 1 inch up.

For example, you might ask me how far I traveled today, and I might answer "4". You should naturally wonder-"four..what?". Four miles, kilometers, city blocks, yards? The standard basis is, effectively, using whatever "unit" $1$ is standing for in the real numbers, and applying the SAME SCALE orthogonally to each dimension. Other scales, and "skewings" are possible, and perhaps more convenient. We get to choose a basis as it suits us.

The points in the plane on the line $y = ax$ have a simpler form if we rotate by the angle $\arctan(a)$. This doesn't change the "line-ness" of the points on the line, but it may lead to less messy calculations. With a lot of linear transformations, finding a "simpler" form of the matrix (by choosing the right basis) is desirable.
 
  • #5
Hi,
Maybe your "blind" spot is geometrical in nature. First think in 2 dimensions (it's easier to draw pictures here). Consider the problem of reflection of points about the x-axis; i.e. a point $(x,y)\rightarrow (x,-y)$. The matrix of this transformation (with respect to the standard basis) is clearly
$$\begin{bmatrix}1&0\\0&-1\end{bmatrix}$$
Now in three dimensions, the problem of reflecting points about the x-y plane is entirely similar. A point $(x,y,z)\rightarrow(x,y,-z)$. The matrix of this transformation, again with respect to the standard basis, is
$$\begin{bmatrix}1&0&0\\0&1&0\\0&0&-1\end{bmatrix}$$
 
  • #6
Thanks guys, always appreciate the patience.

So $ \left[ v_1 \right]_B = (1,0,0) $; is $ v_1 \:also\: = (1,0,0)? $ ie. w.r.t. the standard basis.
 
  • #7
ognik said:
Thanks guys, always appreciate the patience.

So $ \left[ v_1 \right]_B = (1,0,0) $; is $ v_1 \:also\: = (1,0,0)? $ ie. w.r.t. the standard basis.

Only if $v_1 = e_1$. What $v_1$ is in the standard basis will depend, in general, on WHICH plane we are reflecting through, if we were, for example, reflecting in the $xy$-plane, then we would have $v_1 = e_1$.

The point of the approach taken by Khan Academy, is that GIVEN the plane, we can presumably FIND $v_1,v_2$ that span it. Unless the plane contains the $x$-axis as a subspace, $e_1$ (the vector (1,0,0) in the standard basis) will not be one of those spanning vectors.

(N.B. I am using the "usual" identification of $e_1$ with the $x$-axis unit vector (also denoted $\mathbf{i}$), $e_2$ with the $y$-axis unit vector (sim. $\mathbf{j}$), and $e_3$ with the $z$-axis unit vector (aka $\mathbf{k}$). This identification is COMPLETELY ARBITRARY, and should not be taken as gospel).
 
  • #8
Deveno said:
In the basis $B = \{v_1,v_2,v_3\}$ we have $[v_1]_B = (1,0,0)$

Sorry, how do we know that?
 
  • #9
Because:

$v_1 = 1v_2 + 0v_2 + 0v_3$ and since $B$ is a basis, this is the UNIQUE (the only one!) way to write $v_1$ as a linear combination of the $v_j$.

If we had ANY OTHER WAY, say:

$v_1 = av_1 + bv_2 + cv_3$, then:

$0 = v_1 - v_1 = (av_1 + bv_2 + cv_3) - (1v_1 + 0v_2 + 0v_3) = (a-1)v_1 + (b-0)v_2 + (c-0)v_3$

$= (a-1)v_1 + bv_2 + cv_3$.

By the linear independence of $v_1,v_2,v_3$ (since they are a basis), this means:

$a - 1 = b = c = 0$, which means our "two ways" are really both the "same way".
 
  • #10
Thanks all, I believe I have sorted out the muddle in my head - just a temporary inability to see the obvious :-)
 

1. What is change of basis and why is it important?

Change of basis is a mathematical concept used to represent a vector or a linear transformation in a different coordinate system. It is important because it allows us to easily manipulate and analyze mathematical objects in a more convenient basis, and it also helps us to understand the underlying structure and patterns in the data.

2. How do you perform a change of basis?

To perform a change of basis, you need to first identify the original basis and the new basis. Then, you can use a transformation matrix to convert the coordinates of a vector or a linear transformation from the original basis to the new basis. This transformation matrix can be obtained by expressing the basis vectors of the new basis in terms of the basis vectors of the original basis.

3. What are some common applications of change of basis?

Change of basis is commonly used in many fields such as mathematics, physics, and engineering. In mathematics, it is used in linear algebra, differential equations, and functional analysis. In physics, it is used to analyze and solve problems in quantum mechanics, electromagnetism, and fluid dynamics. In engineering, it is used in image processing, signal processing, and control systems.

4. How does change of basis relate to eigenvectors and eigenvalues?

Eigenvectors and eigenvalues play a crucial role in change of basis. The eigenvectors of a linear transformation form a basis for the vector space, and the corresponding eigenvalues represent the scaling factor along each eigenvector. These eigenvectors can be used to construct a transformation matrix, which can then be used to perform a change of basis.

5. What are some common challenges when dealing with change of basis?

Some common challenges when dealing with change of basis include identifying the correct basis to use, finding the appropriate transformation matrix, and understanding how the new basis affects the interpretation of the data or the problem at hand. It may also be challenging to perform a change of basis for complex or high-dimensional data.

Similar threads

  • Linear and Abstract Algebra
Replies
1
Views
828
  • Linear and Abstract Algebra
Replies
3
Views
312
Replies
3
Views
2K
  • Linear and Abstract Algebra
Replies
6
Views
985
Replies
27
Views
1K
  • Linear and Abstract Algebra
Replies
2
Views
915
Replies
3
Views
2K
  • Linear and Abstract Algebra
Replies
3
Views
1K
  • Linear and Abstract Algebra
Replies
6
Views
1K
  • Linear and Abstract Algebra
Replies
1
Views
1K
Back
Top