It is defined that for any matrix A, [itex] A^{0} [/itex] is defined

  • Thread starter Bipolarity
  • Start date
  • Tags
    Matrix
In summary: But, matrix multiplication is non-commutative, so I don't see the benefit of defining A^{0} = I to simplify the matrix binomial theorem, when the binomial theorem is not simple for non-commutative matrices in the first place.
  • #1
Bipolarity
776
2
It is defined that for any matrix A, [itex] A^{0} [/itex] is defined as being equal to the identity matrix, I

My question: Does this definition also hold if A is the zero matrix?

In the real numbers, [itex]0^{0}=1 [/itex] by definition, but I was wondering if this definition extends to matrices

Thanks.

BiP
 
Physics news on Phys.org
  • #2


Bipolarity said:
It is defined that for any matrix A, [itex] A^{0} [/itex] is defined as being equal to the identity matrix, I

My question: Does this definition also hold if A is the zero matrix?

If the definition talks about "any matrix A", then yes, it also holds for the zero matrix.
Where did you find this definition?

In the real numbers, [itex]0^{0}=1 [/itex] by definition,

Not every mathematician accepts this definition, by the way.
 
  • #3


Bipolarity said:
It is defined that for any matrix A, [itex] A^{0} [/itex] is defined as being equal to the identity matrix, I
I think you will find that [itex]A^0= I[/itex] is only defined for A NOT the 0 matrix.

My question: Does this definition also hold if A is the zero matrix?

In the real numbers, [itex]0^{0}=1 [/itex] by definition, but I was wondering if this definition extends to matrices
If that is true "by definition", don't you have a problem with [itex]\lim_{x\to 0} 0^x[/itex]?

[itex]Thanks.

BiP[/QUOTE]
 
  • #4


Bipolarity said:
It is defined that for any matrix A, [itex] A^{0} [/itex] is defined as being equal to the identity matrix, I
Wrong. The matrix needs to be non-singular.
 
  • #5


micromass said:
If the definition talks about "any matrix A", then yes, it also holds for the zero matrix.
Where did you find this definition?

Not every mathematician accepts this definition, by the way.

HallsofIvy said:
I think you will find that [itex]A^0= I[/itex] is only defined for A NOT the 0 matrix.

If that is true "by definition", don't you have a problem with [itex]\lim_{x\to 0} 0^x[/itex]?

Dickfore said:
Wrong. The matrix needs to be non-singular.

I found the matrix definition in Anton's Linear Algebra, in section 1.4 where he defines inverse matrices and proves/defines their properties. In his book he mentions any square matrix A, so I assume he also means 0 matrices since he does not exclude them from the definition. Maybe he did not notice this, so I want to confirm this.

It seems there is contoversy over this. I believe the statement that [itex] 0^{0}=1 [/itex] is verified by Google, then again Google is not a mathematician. I had this question a while ago, but it was clarified by http://www.askamathematician.com/20...ematicians-and-high-school-teachers-disagree/ whose arguments I found quite convincing.

I see no reason why [itex]0^{0}=1[/itex] interferes with [itex]\lim_{x\to 0} 0^x[/itex] since the latter need not be defined at x=0 for the limit to exist. I see no reason why mathematicians would want that function to be continuous at x=0 either...

According to the link above, the definition [itex] 0^{0} = 1 [/itex] allows for a more elegant enunciation of the binomial theorem, among a few other things.

I am confused about the question on matrices. Micro, you seem to say that the 0 exponent of any matrix (including the 0 matrix) is the identity matrix, but Ivy says it is only true for non-zero matrices as bases.

And Dickfore claims it is only true for invertible matrices, I don't know where he's getting this information.

Who is right? Is there any international authority on this subject? For example, in chemistry these conventions would be verified by the IUPAC, I was curious if there is a mathematical authority that can officially define these contentious terms.

Thanks to everyone!

BiP
 
Last edited:
  • #6


Bipolarity said:
It seems there is contoversy over this. I believe the statement that [itex] 0^{0}=1 [/itex] is verified by Google, then again Google is not a mathematician. I had this question a while ago, but it was clarified by http://www.askamathematician.com/20...ematicians-and-high-school-teachers-disagree/ whose arguments I found quite convincing.

I see no reason why [itex]0^{0}=1[/itex] interferes with [itex]\lim_{x\to 0} 0^x[/itex] since the latter need not be defined at x=0 for the limit to exist. I see no reason why mathemat
icians would want that function to be continuous at x=0 either...

According to the link above, the definition [itex] 0^{0} = 1 [/itex] allows for a more elegant enunciation of the binomial theorem, among a few other things.

The binomial theorem relies implicitly on the commutativity of multiplication of ordinary number. That is why there is a combinatorial coefficient in front of [itex]a^{n - k} b^{k}[/itex], because you clump all products of the type, say:
[tex]
a b a b a, \ a a b b a, a b a a b, \ldots
[/tex]
together.

But, matrix multiplication is non-commutative, so I don't see the benefit of defining [itex]A^{0} = I[/itex] to simplify the matrix binomial theorem, when the binomial theorem is not simple for non-commutative matrices in the first place.

Bipolarity said:
And Dickfore claims it is only true for invertible matrices, I don't know where he's getting this information.

I am starting from a general definition of a function of a (square) matrix. Namely, suppose we can diagonalize a (square) matrix:
[tex]
A = U \, \Lambda \, U^{-1}
[/tex]
where [itex]\Lambda[/itex] is a diagonal matrix containing the eigenvalues of A on the main diagonal. Then, one defines:
[tex]
f(A) = U \, f(\Lambda) \, U^{-1}
[/tex]
where [itex]f(\Lambda)[/itex] is simply a diagonal matrix again whose diagonal elements are the values [itex]f(\lambda_\alpha)[/itex] for each corresponding eigenvalue of A, and U is the same matrix as above.

Then, the problem of defining [itex]A^0[/itex] for a matrix goes down to the problem for defining it for a (complex) number [itex]\lambda^0[/itex]. Taking powers of complex numbers is given by:
[tex]
u^{v} = \exp \left(v \, \mathrm{Log}(u) \right)
[/tex]
If you take [itex]v = 0[/itex], it seems the argument of the exponential is always zero, and [itex]\exp(0) = 1[/itex]. But, that is true provided that [itex]\mathrm{Log}(u)[/itex] exists! Otherwise the r.h.s. is not defined. And, Log is not defined for [itex]u = 0[/itex].

If a matrix has a zero eigenvalue, then and only then its determinant, being the product of its eigenvalues is zero as well, and the matrix is singular. This is why I required the determinant to be non-zero, so that you have only non-zero eigenvalues, and their exponentiation (even by 0) is defined.
 
Last edited:
  • #7


Dickfore said:
But, matrix multiplication is non-commutative, so I don't see the benefit of defining [itex]A^{0} = I[/itex] to simplify the matrix binomial theorem, when the binomial theorem is not simple for non-commutative matrices in the first place.
That non-commutativity of matrix multiplication is a bit of a red herring here. While matrix multiplication is not commutative in general, a positive power of some matrix [itex]A[/itex] does commute with the matrix [itex]A[/itex].

The reason for denoting [itex]A^0=I[/itex] for any square matrix (singular or not) is the same reason one denotes [itex]x^0=1[/itex]. It's an extremely convenient abuse of notation.

Strictly speaking, expressing some function [itex]f(x)[/itex] as a power series [itex]f(x) = \sum_{n=0}^{\infty} a_n x^n[/itex] is incorrect because of the [itex]0^0[/itex] problem. To be pedantically correct, that power series should be written as [itex]f(x) = a_0 + \sum_{n=1}^{\infty} a_n x^n[/itex]. Denoting [itex]0^0=1[/itex] as a convenient abuse of notation is what let's us get away with the more compact representation.

The same applies to matrices. The exponential of a square matrix [itex]A[/itex] is typically written as [itex]\exp(A) = \sum_{n=0}^{\infty} \frac 1 {n!} A^n[/itex]. The pedantically correct version is [itex]\exp(A) = I + \sum_{n=1}^{\infty} \frac 1 {n!} A^n[/itex]. The compact representation arises by denoting [itex]A^0=I[/itex] for all square matrices [itex]A[/itex], including singular ones. It's an abuse of notation, but a very convenient one.
 
  • #8


I too think we should accept ##0^0=1## (also for zero matrices). It is true that ##0^x## becomes discontionuous at ##0## then, but this is only a minor inconvenience, IMHO.

My main reason for this is that it is logical for natural numbers, since, for positive integers ##m## and ##n##, ##m^n## is the number of functions from a set with ##n## elements to a set with ##m## elements. If we want this to hold also for ##m=0## and/or ##n=0##, we must put ##m^0=1## for all natural numbers ##m##, including ##m=0##, and ##0^n=0## for all natural numbers ##n## except ##n=0##. So, ##0^0=1##.

What do I mean?

Recall that a function ##f## from set ##A## to a set ##B## (more precisely, the graph of such a function), is a subset of the cartesian product ##A\times B## such that to each ##a\in A##, there is exactly one ##b\in B## such that ##(a,b)\in f##. We then write ##f(a)=b## instead of ##(a,b)\in f##.
It is easy to see that there are ##m^n## functions from ##A## to ##B##, if ##A## has ##n## elements and ##B## has ##m## elements, and ##m## and ##n## are positive integers.

Now, if ##n=0##, that is, if ##A## is the empty set, ##\varnothing##, then ##A\times B=\varnothing\times B=\varnothing##. This set has only one subset: ##\varnothing## itself. But the condition "For each ##a\in A=\varnothing## there is exactly one ##b\in B## such that ##(a,b)\in f=\varnothing##" is vacuously satisfied for ##f=\varnothing##, since there are no ##a\in A=\varnothing##. This means that ##f=\varnothing## is (the graph of) a function from ##A=\varnothing## to ##B##, and it is the only (graph of a) function of this kind, since ##A\times B=\varnothing## has no more subsets.
Thus, there is exactly one function from a set ##A## with ##0## elements to a set ##B## with ##m## elements. It is thus logical to set ##m^0=1## for all natural numbers ##m##.
Notice that the argument above is valid also for ##m=0##, that is, for ##B=\varnothing##. Consequently, we set ##0^0=1##.

Assume instead that ##m=0## and ##n>0##. Then ##A\neq\varnothing## and ##B=\varnothing##. Again, we have ##A\times B=A\times\varnothing=\varnothing##, and this has only one subset, ##f=\varnothing## itself. But in this case, the condition "For each ##a\in A=\varnothing## there is exactly one ##b\in B## such that ##(a,b)\in f=\varnothing##" is not satisfied, since there are elements ##a\in A## but no elements ##(a,b)\in f=\varnothing##. ##f=\varnothing## is therefore not a (graph of a) function from ##A## to ##B=\varnothing##. Since ##A\times B=\varnothing## has no more subsets, there are no other functions of this kind either.
Thus, there are no functions from ##A## to ##B=\varnothing##. It is therefore logical to set ##0^n=0## for all natural numbers ##n>0##.
 
Last edited:
  • #9


Bipolarity, as you can see, there is no agreement about [itex]0^0[/itex] among mathematicians. All in all, it is just a definition, we could define [itex]0^0=10[/itex] if we want, it wouldn't change much (but it's not like anybody defines it like that).

So, the definition of [itex]0^0[/itex] (as matrices or as numbers), depends crucially on the author. Anton seems to use the definition [itex]0^0=1[/itex]. There is nothing wrong with that.
Other authors might leave it undefined. You have to check with each author and see what he says about it.

So there is no right and wrong answer here. It is merely a definition. So pick the definition that you like best.
 
  • #10


micromass said:
Bipolarity, as you can see, there is no agreement about [itex]0^0[/itex] among mathematicians. All in all, it is just a definition, we could define [itex]0^0=10[/itex] if we want, it wouldn't change much (but it's not like anybody defines it like that).

So, the definition of [itex]0^0[/itex] (as matrices or as numbers), depends crucially on the author. Anton seems to use the definition [itex]0^0=1[/itex]. There is nothing wrong with that.
Other authors might leave it undefined. You have to check with each author and see what he says about it.

So there is no right and wrong answer here. It is merely a definition. So pick the definition that you like best.

Thanks!

I've decided to define [itex] 0^{0} = π + e + 1 [/itex] so to make the binomial theorem complicated for future generations!

Just kidding.

BiP
 
  • #11


"It is true that 0x becomes discontionuous at 0 then, but this is only a minor inconvenience, IMHO."

Well, what do you think the limit of x^(k(x)/ln(x)) is, as x goes to 0, the limit of k(x) as x goes to zero being finite?
 
Last edited:
  • #12


arildno said:
"It is true that 0x becomes discontionuous at 0 then, but this is only a minor inconvenience, IMHO."

Well, what do you think the limit of x^(k/ln(x)) is, as x goes to 0?
##e^k##. So...?
 
  • #13


Erland said:
##e^k##. So...?
Meaning it is as silly and nonsensical to define 0^0 to have some value, as it is silly to define 0/0 to have some value.
 
  • #14


arildno said:
Meaning it is as silly and nonsensical to define 0^0 to have some value, as it is silly to define 0/0 to have some value.

Uuuh, how does that follow from the limit being [itex]e^k[/itex]?

It makes a lot of sense to define [itex]0^0=1[/itex]. A lot of mathematicians make that definition. Saying it is silly and nonsensical is really not justified.
 
  • #15


arildno said:
Meaning it is as silly and nonsensical to define 0^0 to have some value, as it is silly to define 0/0 to have some value.
We have seen some examples where the definition ##0^0=1## makes sense and is useful. That suffices for me. I have not seen any examples where it is paticularly meaningful to define 0/0 as a particular value.
 
  • #16


micromass said:
Uuuh, how does that follow from the limit being [itex]e^k[/itex]?

It makes a lot of sense to define [itex]0^0=1[/itex]. A lot of mathematicians make that definition. Saying it is silly and nonsensical is really not justified.

You might as well define it to be 2.
You might even find some uses for that as well.
 
  • #17


arildno said:
You might as well define it to be 2.

Sure.

You might even find some uses for that as well.

I doubt you'll find much uses. Defining [itex]0^0=1[/itex] is really useful and it simplifies a lot of formulas. It is obvious that it is much more useful than defining it to be 2.
 
  • #18


micromass said:
Sure.



I doubt you'll find much uses. Defining [itex]0^0=1[/itex] is really useful and it simplifies a lot of formulas. It is obvious that it is much more useful than defining it to be 2.

And in those formulas, you evidently provide the EXPLICIT definition of 0^0 being..1, thus making it a particular case valid for that formula, rather than any general result or definition.
 
  • #19


arildno said:
And in those formulas, you evidently provide the EXPLICIT definition of 0^0 being..1, thus making it a particular case valid for that formula, rather than any general result or definition.

When somebody writes down the binomial theorem, then I have never seen him say explicitely that we assume [itex]0^0=1[/itex].

But in general, you certainly do make a general definition.
If [itex]\kappa[/itex] and [itex]\kappa^\prime[/itex] are cardinal numbers and if X is a set with cardinality [itex]\kappa[/itex] and Y has cardinality [itex]\kappa^\prime[/itex]. Then we say by definition that [itex]X^Y[/itex] has cardinality [itex]\kappa^{\kappa^\prime}[/itex].

If you take in that definition [itex]\kappa=\kappa^\prime=0[/itex]. Then by definition, the cardinal [itex]0^0[/itex] is the size of the set [itex]\emptyset^\emptyset[/itex]. This size is one.
So at least for cardinal numbers, we do make the explicit and general definition [itex]0^0=1[/itex].
 

1. What is the definition of [itex] A^{0} [/itex] for a matrix A?

The definition of [itex] A^{0} [/itex] for a matrix A is the identity matrix, denoted as [itex] I [/itex]. It is a square matrix with 1s along the main diagonal and 0s elsewhere.

2. Why is [itex] A^{0} [/itex] defined as the identity matrix?

[itex] A^{0} [/itex] is defined as the identity matrix because it is the only matrix that when multiplied with any other matrix A, results in the original matrix A. In other words, [itex] A^{0}A = A [/itex]. This property is important in matrix operations and simplifying calculations.

3. Is [itex] A^{0} [/itex] always defined for any matrix A?

Yes, [itex] A^{0} [/itex] is always defined for any matrix A. The definition of [itex] A^{0} [/itex] as the identity matrix applies to all square matrices. However, it is important to note that the dimensions of the identity matrix [itex] I [/itex] must match the dimensions of matrix A.

4. How does raising a matrix to the 0th power affect its elements?

Raising a matrix A to the 0th power does not change its elements. The resulting matrix is still the original matrix A, as shown by [itex] A^{0}A = A [/itex]. Therefore, all elements in [itex] A^{0} [/itex] remain the same as those in matrix A.

5. Can [itex] A^{0} [/itex] be used in matrix calculations?

Yes, [itex] A^{0} [/itex] can be used in matrix calculations. It can be used to simplify calculations by replacing [itex] A^{0} [/itex] with the identity matrix [itex] I [/itex]. It can also be used in matrix exponentiation, where [itex] A^{n} [/itex] is calculated by multiplying A with itself n times, starting with [itex] A^{0} = I [/itex].

Similar threads

  • Linear and Abstract Algebra
Replies
8
Views
993
  • Linear and Abstract Algebra
Replies
1
Views
764
  • Linear and Abstract Algebra
Replies
6
Views
405
  • Linear and Abstract Algebra
Replies
12
Views
1K
  • Linear and Abstract Algebra
Replies
2
Views
493
Replies
7
Views
782
  • Linear and Abstract Algebra
Replies
6
Views
1K
  • Linear and Abstract Algebra
Replies
1
Views
977
  • Linear and Abstract Algebra
Replies
8
Views
662
  • Linear and Abstract Algebra
Replies
7
Views
2K
Back
Top