Why do we need to convert to a diagonal matrix?

matqkks
Messages
280
Reaction score
5
Apart from simplifying matrix powers, why do we want to diagonalize a matrix? Do they have any appealing application which can be used to motivate to study diagonal matrices.
Thanks for any answers.
 
Physics news on Phys.org
Diagonal matrices are nice because ANY matrix calculation is simpler with diagonal matrices. For example, any two diagonal matrices compute. Also, it is trivial to find the inverse of a diagonal matrix. Apart from these facts and the fact that numerical computations are easier and much more stable, there are theoretical reasons to want to deal with diagonal matrices. Proving things about diagonal matrices is quite a bit easier than proving things about general matrices.
 
One example in which diagonalization is important is the study of quadratic forms.
http://en.wikipedia.org/wiki/Quadratic_form

A quadratic form can be written as
Q(x) = x^TAx
where x is a column vector and A is a symmetric matrix. It is a theorem that you can always diagonalize A by a rotation of coordinates. For example, in 2D, if you have an equation such as: ax^2+2bxy+cy^2 = D, then by rotating your coordinate axes you can rewrite the equation as
A\bar{x}^2+B\bar{y}^2 = D
in your new coordinates. Therefore, the original equation represents an ellipse or a hyperbola (or possibly a pair of parallel lines if one of the eigenvalues A or B is zero.)

Quadratic forms are important, for example, because a general function f(x,y,z) has a local Taylor polynomial approximation
f = f(P) + df + Q_f + higher order terms
The second order term is a quadratic form which is determined by the Hessian matrix. So, for example, at a critical point (where the differential df =0), the first nonzero term in \Delta f is the quadratic form determined by the Hessian. Since all quadratic forms can be diagonalized by a rotation of coordinates, that means that by a rotation of coordinates,
\Delta f = A\bar{x}^2+B\bar{y}^2+C\bar{z}^2 + higher order terms
A, B, C are the eigenvalues of the Hessian. One thing you can do with this knowledge is determine whether a critical point is a maximum. To do that you check the eigenvalues of the Hessian matrix. If they are all negative, then you have a relative maximum.
 
A good application is in the study of 2nd order differential equations.
This can be seen in Structural dynamics

Equations can arise in the form of

My'' + y' + Ky = F

where M,C,K are (nxn) matrices and y'',y',y, and F are (nx1) vectors

The n equations are coupled with each other. If we can diagonalize M,C,K then we uncouple them and we can then solve n independent equations.

for example, let's assume there is a matrix \Phi such that

\Phi^{T} M \Phi = M (Diagonal)
\Phi^{T} C \Phi = C (Diagonal)
\Phi^{T} K \Phi = K (Diagonal)

Then if we let y = \Phiu then

M\Phiu'' + C\Phiu' + K\Phiu = F

Multiply by transpose \Phi^{T} to get

\Phi^{T}M\Phiu'' + \Phi^{T}C\Phiu' + \Phi^{T}K\Phiu = \Phi^{T}F

which simplifies to

Mu'' + Cu' + Ku = \Phi^{T}F
which is just n independent equations which can be solved separately to find each component in the vector u.

Once that is done, the vector y can be found by y = \Phiu

I hope this helps and I hope this was readable.
 
Diagonalizing matrices can help computer run-times as well.
 
Thread 'Determine whether ##125## is a unit in ##\mathbb{Z_471}##'
This is the question, I understand the concept, in ##\mathbb{Z_n}## an element is a is a unit if and only if gcd( a,n) =1. My understanding of backwards substitution, ... i have using Euclidean algorithm, ##471 = 3⋅121 + 108## ##121 = 1⋅108 + 13## ##108 =8⋅13+4## ##13=3⋅4+1## ##4=4⋅1+0## using back-substitution, ##1=13-3⋅4## ##=(121-1⋅108)-3(108-8⋅13)## ... ##= 121-(471-3⋅121)-3⋅471+9⋅121+24⋅121-24(471-3⋅121## ##=121-471+3⋅121-3⋅471+9⋅121+24⋅121-24⋅471+72⋅121##...
##\textbf{Exercise 10}:## I came across the following solution online: Questions: 1. When the author states in "that ring (not sure if he is referring to ##R## or ##R/\mathfrak{p}##, but I am guessing the later) ##x_n x_{n+1}=0## for all odd $n$ and ##x_{n+1}## is invertible, so that ##x_n=0##" 2. How does ##x_nx_{n+1}=0## implies that ##x_{n+1}## is invertible and ##x_n=0##. I mean if the quotient ring ##R/\mathfrak{p}## is an integral domain, and ##x_{n+1}## is invertible then...
The following are taken from the two sources, 1) from this online page and the book An Introduction to Module Theory by: Ibrahim Assem, Flavio U. Coelho. In the Abelian Categories chapter in the module theory text on page 157, right after presenting IV.2.21 Definition, the authors states "Image and coimage may or may not exist, but if they do, then they are unique up to isomorphism (because so are kernels and cokernels). Also in the reference url page above, the authors present two...
Back
Top