# Linear transformation

## Main Question or Discussion Point

How important are linear transformations in linear algebra? In some texts linear transformations are introduced first and then the idea of a matrix. In other books linear transformations are relegated to being an application of matrices. What is the best way of introducing linear transformation on a linear algebra course? How do we motivate students to study transformations as part of linear algebra? What is their real impact?

Related Linear and Abstract Algebra News on Phys.org
I personally feel that linear transformations can and should be taught without matrices at all. Matrices are a means of representing such transformations; they're not the transformations themselves. While students should be familiar with matrices as a matter of practicality, as a physicist, I can see the geometric content of this...

$$\underline P(a) = a - (a \cdot \hat n) \hat n$$

...more than I can in just being given the matrix $\underline P(e_i) \cdot e_j$. But I know math is much, much bigger and broader than what I as a physicist use, so perhaps some of that breadth makes the matrix representation necessary.

Last edited:
mathwonk
Homework Helper
the basic example of a linear transformation in mathematics is perhaps differentiation.

this is not readily described by a matrix since the space of differentiable functions is infinite dimensional.

however when restricted to a finite dimensional subspace such as polynomials of degree ≤ n, one obtains the basic example of "nilpotent"matrix.

when restricted to a finite dimensional space of functions of form (e^ct)t^k, for 0≤ k ≤ n, one obtains as matrix a basic jordan block.

failure to give this example early on may explain why students of advanced linear algebra find jordan form so strange.

for a presentation of linear algebra including these examples, you may see either the nice book by Insel, Friedberg, and Spence, or my free notes (#7) for math 4050 on my website at UGA.

http://www.math.uga.edu/%7Eroy/4050sum08.pdf [Broken]

One difference between these two sources is my heavier reliance on the basic concept of the minimal polynomial for a linear transformation. Their book, replete with examples, is also about 4 times as long as mine.

by the way the famous mathematician Emil Artin, argues in his book Geometric Algebra precisely that linear transformations should be the primary object of study and that matrices should be left out almost entirely, except say when a computation needs to be made, say of a determinant. Then he says to immediately throw them out again afterwards. I.e. matrices are entirely a device for representing and computing with linear transformations.

A matrix is given by a pair: (linear transformation of a vector space, basis for that vector space)

equivalently, given a vector space and a linear transformation, matrices for that transformation are equivalent to bases of the vector space. Hence a good choice of basis may be expected give a good matrix.

Last edited by a moderator:
mathwonk
Homework Helper
oops my bad. i wasn't thinking. several different bases may give the same matrix, so the two concepts are not equivalent. e.g. the zero transformation has the same matrix in all bases, as does the identity transformation. a basis gives a matrix. also given a basis, transformations are equivalent to matrices.

the concept sliding around in my memory was that given an n dimensional vector space V, bases of V are equivalent to isomorphisms of V with k^n.

Fredrik
Staff Emeritus
Gold Member
I think it's better to introduce linear transformations first. Matrices can be introduced as soon as the student understand bases. It might be a good idea to first use the notation that puts the row index upstairs, because that makes it almost impossible to get the formula for the components of a linear transformation with respect to a pair of bases wrong: $A^i_j=(Ae_j)^i$. This way the definition of matrix multiplication can be properly motivated. If $A:Y\to Z$ and $B:X\to Y$ are linear transformations, and [A] and denote the matrices corresponding to A and B (given bases for X,Y,Z), the product [A] is defined as [A$\circ$B], i.e. as the matrix corresponding to $A\circ B:X\to Z$.

Last edited by a moderator: