Row/column operation on matrices and determinants

In summary, simultaneous row and column operations cannot be applied on a matrix when finding its inverse by elementary transformation. However, they can be applied in the determinant case. The kernel and image may be disturbed in a matrix, but this does not occur in the determinant case. Row operations preserve the kernel, while column operations preserve the image. However, in certain cases, such as with invertible matrices, the image and kernel may be "preserved" by both row and column operations.
  • #1
Raghav Gupta
1,011
76
How we cannot apply row and column operation simultaneously on matrix when finding its inverse by elementary transformation but can apply it in determinant?
I think kernel and image gets disturbed in a matrix, though I don't know what it actually is.
Why not in determinant case?
 
Physics news on Phys.org
  • #2
Raghav Gupta said:
How we cannot apply row and column operation simultaneously on matrix when finding its inverse by elementary transformation
You can - but only when you do it on the right hand side simultaneously. Assume you want to find [itex]B [/itex] such that [itex]A \cdot B = I [/itex]. You do the same row operations on [itex] A[/itex] and [itex]I [/itex]. After a while (excluding singularities etc.) you end up with [itex] U \cdot B = R[/itex], where [itex] U[/itex] is upper triangular. Now substitute back until you end up with [itex]I \cdot B = H [/itex] and you are done.
 
  • #3
But doesn't the kernel or image gets disturbed?
 
  • #4
Raghav Gupta said:
But doesn't the kernel or image gets disturbed?

I am not an expert on matrices so, I cannot answer that. What I know, is how to find the inverse (as detailed above). Useful facts:
  • The matrix can be inverted if its determinant is different from 0.
  • The determinant of the inverse matrix is the inverse of the determinant of the original matrix.
See also http://en.wikipedia.org/wiki/Invertible_matrix .
 
  • #5
Raghav Gupta said:
How we cannot apply row and column operation simultaneously on matrix when finding its inverse by elementary transformation but can apply it in determinant?
I think kernel and image gets disturbed in a matrix, though I don't know what it actually is.
Why not in determinant case?

Row operations are equivalent to left multiplication by corresponding elementary matrices, column operations are equivalent to right multiplication. So, when you perform row operations on a (square) matrix ##A##, you get a matrix ##E## (the product of corresponding elementary matrices) such that ##EA=I##. For square matrices that means that ##E=A^{-1}##; performing the same row operations on the right hand side ##I## you get ##EI=E=A^{-1}##.

Again, you can find the inverse performing only column operations: you get a matrix ##E## (the product of corresponding elementary matrices) such that ##AE=I##, which again means that ##E=A^{-1}##. Performing the same column operations on the right hand side ##I## you get there ##IE=E=A^{-1}##.

If you perform both row and column operations you get two matrices ##E_1## and ##E_2## such that ##E_1AE_2=I##, so in the right hand side you get ##E_1IE_2=E_1E_2##. For ##E_1E_2## to be the inverse of ##A## you need ##AE_1E_2=I## (or equivalently ##E_1E_2A =I##), and you have ##E_1AE_2=I##. Matrix multiplication is not commutative, so there is no reason for the latter identity imply the former.

"No reason" is not a formal proof that the method does not work, for the formal proof you need a counterexample. You can find it just by playing with applying row and column operations to ##2\times 2## matrices. There are also more scientific method of constructing a counterexample.

The reason that applying both row and column operations is that while ##E_1AE_2=I## does not imply that ##E_1E_2=A^{-1}## it implies that ##\operatorname{det} (E_1E_2)=\operatorname{det} A^{-1}##. Namely, $$1=\operatorname{det}(E_1AE_1) = \operatorname{det} E_1 \operatorname{det} A \operatorname{det}E_2,$$ so $$\operatorname{det}(E_1E_2) \operatorname{det}E_1\operatorname{det}E_2= (\operatorname{det}A)^{-1} =\operatorname{det}A^{-1}.$$

Finally, for invertible matrices the image is all ##\mathbb R^n## and the kernel is always trivial (i.e. ##\{\mathbf 0\}##), they cannot be "disturbed" by row/column operations, they will remain the same.
 
  • #6
Raghav Gupta said:
But doesn't the kernel or image gets disturbed?
If you're talking about a matrix you get by applying a row operation to another matrix, the answer is no. If you start with a matrix A, and apply one of the three row operations to it to get A1, the matrices A and A1 are equivalent. They have exactly the same kernel and image.
Edit: The image can change. See my later post in this thread.

If you're talking about applying column operations, I don't know -- I have never needed to apply column operations to reduce a matrix. However, if you swap the columns of a matrix, you are swapping the roles of the variables these columns represent.
 
Last edited:
  • #7
Mark44 said:
If you're talking about a matrix you get by applying a row operation to another matrix, the answer is no. If you start with a matrix A, and apply one of the three row operations to it to get A1, the matrices A and A1 are equivalent. They have exactly the same kernel and image.

That is not true. Row operations preserve kernel, column operations preserve image (column space).

In the case of invertible matrices, however, the image is always all ##\mathbb R^n## and the kernel is always ##\{\mathbf 0\}##, so we can say that in this case the image and kernel are "preserved" user row and column operations.
 
  • #8
A bit of specification. Row operations preserve kernel (but generally not image), column operations preserve image (but generally not kernel).
 
  • #9
Hawkeye18 said:
That is not true. Row operations preserve kernel, column operations preserve image (column space).
Hawkeye18 said:
A bit of specification. Row operations preserve kernel (but generally not image), column operations preserve image (but generally not kernel).
I mispoke. Row operations don't necessarily preserve the range, as I said. A simple example shows this:
$$A = \begin{bmatrix} 1 & 0 \\ 0 & 1 \\ 1 & 1\end{bmatrix}$$
Using row reduction, we get an equivalent matrix.
$$B = \begin{bmatrix} 1 & 0 \\ 0 & 1 \\ 0 & 0\end{bmatrix}$$
Although the dimensions of the column spaces of A and B are equal (2), they span different subspaces of R3.

The columns of A represent a plane in R3 that is perpendicular to <-1, -1, 1>. The columns of B represent a different plane in R3 that is perpendicular to the z-axis.
 
  • #10
Mark44 said:
I mispoke. Row operations don't necessarily preserve the range, as I said. A simple example shows this:
$$A = \begin{bmatrix} 1 & 0 \\ 0 & 1 \\ 1 & 1\end{bmatrix}$$
Using row reduction, we get an equivalent matrix.
$$B = \begin{bmatrix} 1 & 0 \\ 0 & 1 \\ 0 & 0\end{bmatrix}$$
I see you have applied R3--------> R3 - R1
But there should be 1 in row 3 column 2 or more simply B32 should be equal to 1?
 
  • #11
Raghav Gupta said:
I see you have applied R3--------> R3 - R1
But there should be 1 in row 3 column 2 or more simply B32 should be equal to 1?
No. Here's what I did: -R1 + R3 --> R3 and then -R2 + R3 --> R3.

IOW, add -R1 to R3, and then add -R2 to R3.
 
  • #12
Mark44 said:
No. Here's what I did: -R1 + R3 --> R3 and then -R2 + R3 --> R3.

IOW, add -R1 to R3, and then add -R2 to R3.

Got it. Thanks to all of you - HawkEye 18, Mark44 and Svein.
 

What are row/column operations on matrices?

Row/column operations on matrices involve manipulating the rows and columns of a matrix in order to make it easier to solve equations or find certain properties of the matrix, such as its determinant.

What are the three main types of row/column operations?

The three main types of row/column operations are: multiplying a row or column by a scalar, interchanging two rows or columns, and adding a multiple of one row or column to another row or column.

How do row/column operations affect the determinant of a matrix?

Row/column operations do not change the value of the determinant, but they can make it easier to calculate. For example, multiplying a row by a scalar will multiply the determinant by the same scalar, while interchanging two rows will cause the determinant to be multiplied by -1.

What is the purpose of using row/column operations on a matrix?

The purpose of using row/column operations is to simplify the matrix and make it easier to solve equations or find certain properties. These operations can also be used to manipulate the matrix into a specific form, such as row echelon form or reduced row echelon form.

What are some real-world applications of row/column operations on matrices and determinants?

Row/column operations and determinants have various applications in fields such as engineering, economics, and computer science. For example, they can be used to solve systems of linear equations, calculate areas and volumes, and determine the stability of a system.

Similar threads

  • Linear and Abstract Algebra
Replies
7
Views
1K
  • Linear and Abstract Algebra
Replies
12
Views
1K
  • Linear and Abstract Algebra
Replies
1
Views
761
  • Linear and Abstract Algebra
Replies
8
Views
886
Replies
13
Views
2K
  • Linear and Abstract Algebra
2
Replies
48
Views
5K
  • Linear and Abstract Algebra
Replies
9
Views
896
  • Precalculus Mathematics Homework Help
Replies
25
Views
988
  • Linear and Abstract Algebra
Replies
4
Views
7K
  • Linear and Abstract Algebra
Replies
2
Views
930
Back
Top