Why Use Matrices for Linear Transformations?

Click For Summary
Using matrices for linear transformations simplifies the process of performing successive transformations, such as rotations, through matrix multiplication. Matrices allow for easier computation of eigenvalues and eigenvectors, providing insights like the axis of rotation without losing information. They convert conceptual operations into numerical ones, making calculations more straightforward, especially in physical applications with predefined coordinate systems. For example, computing derivatives using matrix representations clarifies operations across different bases. Overall, matrices enhance the efficiency and transparency of linear transformations in various contexts.
matqkks
Messages
282
Reaction score
5
Why would you want to use a matrix for a linear transformation?
Why not just use the given transformation instead of writing it as a matrix?
 
Physics news on Phys.org
If you have a number of rotations to be performed in succession, you can just multiply the matrices. Also you can determine information about a rotation, for example the axis of rotation, by calclulating the eigevectors of the matrix.
 
Using the transformation or using the matrix is equivalent. You won't lose information if you use the matrix.

If you want to keep on using the transformation, then you can do this. But in many cases, using the matrix is simply much easier. Finding eigenvalues for example is much easier with a matrix than with a transformation.
 
one reason is:

a matrix calculation reduces the computation of composition of linear transformations, as well as the computation of image elements under a linear transformation, to arithmatic operations in the underlying field. that is:

conceptual--->numerical.

sometimes, this is preferrable for getting "actual answers" in a physical application, where some preferred basis (coordinate system) might already be supplied.

for example, the differentiation operator is a linear transformation from Pn(F) to Pn(F).

actually "computing a derivative" IS just computing the matrix product [D]B[p]B = [p']B:

for n = 2, and F = R, we have for the basis B = {1,x,x2}, that [D]B=

[0 1 0]
[0 0 2]
[0 0 0],

or that if p(x) = a + bx + cx2,

p'(x) = b + 2cx.

of course, this would be just as easy using D(p) = p' using the calculus definition,

but it's not so clear what happens if you want to use THIS basis: {1+x,1-x,1-x2}, using the calculus definition, whereas the matrix form makes it transparent.
 
Is it by using a matrix representation of a derivative that CAS and programmable calculators evaluate derivatives?
 
Thread 'How to define a vector field?'
Hello! In one book I saw that function ##V## of 3 variables ##V_x, V_y, V_z## (vector field in 3D) can be decomposed in a Taylor series without higher-order terms (partial derivative of second power and higher) at point ##(0,0,0)## such way: I think so: higher-order terms can be neglected because partial derivative of second power and higher are equal to 0. Is this true? And how to define vector field correctly for this case? (In the book I found nothing and my attempt was wrong...

Similar threads

  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 19 ·
Replies
19
Views
4K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 12 ·
Replies
12
Views
5K
  • · Replies 9 ·
Replies
9
Views
2K
Replies
2
Views
2K