Why is 0! equal to 1 in mathematics?

  • Thread starter Thread starter murshid_islam
  • Start date Start date
  • Tags Tags
    Mystery
AI Thread Summary
The discussion centers on the mathematical definition of 0!, which is defined as equal to 1 for consistency in factorial calculations. This definition allows for the recursive relationship n! = n * (n-1)!, maintaining continuity when n equals 1. The reasoning includes the idea that there is one way to arrange zero objects, supporting the notion that 0! should equal 1. Additionally, it is noted that the factorial function can be defined independently of its initial interpretation related to counting arrangements. Ultimately, the consensus is that defining 0! as 1 is a practical and logical choice in mathematics.
murshid_islam
Messages
468
Reaction score
21
can anyone please help me with this. why is 0! = 1? :confused:

thanks in advance to anyone who can help.
 
Mathematics news on Phys.org
Because that is what we define it to be. There are hundreds of posts on this topic.

We could choose not to define it at all and state that n! is the number of ways or ordering n things for integer n strictly positive, and not define 0!. But that turns out to be unwieldy when doing things like nCr, so for ease, consistency, whatever, there is no harm in defining 0!=1. End of story, no great mystery here.
 
Last edited:
so mathematicians just decided that 0! = 1? is that what you mean?
 
murshid_islam said:
so mathematicians just decided that 0! = 1? is that what you mean?
It didn't have any prior meaning before the mathematicians appropriated it, so yes, they were free to decide what that symbol collection 0! should mean.

The conventional definition of the factorial goes like this:
0!=1, n!=n*((n-1)!), n\geq{1}
where n is always a natural number.
 
Last edited:
for example, 3! is the product of the first three integers. but 0! does not have such meaning. it is just defined to be 1. am i correct?
 
murshid_islam said:
for example, 3! is the product of the first three integers. but 0! does not have such meaning. it is just defined to be 1. am i correct?
Quite so.:smile:
Or, rather, 3! is recursively defined in such a manner that you can compute it by multiplying together the first three natural numbers.
 
thank you very much for your help, arildno.
 
It's not entiely arbitrary. n!/n = (n-1)! whenever division by n is defined. If we assign any value to 0! other than 1, this would not hold.
 
But a decision to require n!/n = (n-1)! to be true where n=1 is arbitrary.
 
  • #10
Besides, set 0!=a.
Then, by induction, we have:
n!=a*1*2*...(n-1)*n, that is, n!/n=(n-1)! whatever value you assign to 0!.
 
  • #11
shmoe said:
But a decision to require n!/n = (n-1)! to be true where n=1 is arbitrary.

Sure, as is the decision to make 0! the number of ways to arrange 0 objects (1 way). The point is that it is consistent with the way it works for other numbers.
 
  • #12
At first I thought this was a totally absurd question, because in many programming languages "!=" means "not equal to". Of course 0 is not equal to 1

I've been programming all weekend so everything else goes out the door.
 
  • #13
Spiderman said:
At first I thought this was a totally absurd question, because in many programming languages "!=" means "not equal to". Of course 0 is not equal to 1

I've been programming all weekend so everything else goes out the door.
:smile: :smile: :smile:
 
  • #14
The explanation my math teacher gave me was the following:

5!=5*4*3*2*1
4![/color]=4*3*2*1

So

5!=5*4!
4!=4*3!
3!=3*2!
2!=2*1!
1!=1*0!

1!=1...so 1*0! has to equal 1. 1*0!=1 => 0!=1/1 => 0!=1
 
  • #15
Then why isn't 0! 0*(-1)! which must be 0 if (-1)! is defined?
 
  • #16
0! =\Gamma(1) =\int_{0}^{\infty} e^{-t} dt= 1

Daniel.
 
  • #17
I think any thread with the : “0”, “1”, and “=” is asking for trouble.
 
  • #18
0! doesn't have to be defined exactly, look at this relation

\frac{n!}{k!} = (n - k)!

If n = k then you get

\frac{n!}{n!} = (n - n)!
 
  • #19
waht said:
0! doesn't have to be defined exactly, look at this relation

\frac{n!}{k!} = (n - k)!

This isn't true. 5!/4!=5, not (5-4)!=1.
 
  • #20
yea you are right, my bad.
 
  • #21
matt grime said:
Then why isn't 0! 0*(-1)! which must be 0 if (-1)! is defined?

I'm way over my head to answer that, but I've never seen a ! of a negative...the ! starts at 1...

a!=a(a-1)(a-2)...*1

It never goes below 1 and it never goes decimal, so it makes me assume 0! has to be 1*whatever is bigger than 1 and less than zero. or 1 itself...you can *1 it 1000 times if you feel like it.

I'm seing now a pattern...
5! has 5 factors.
4! has 4 factors.
3! has 3 factors.
2! has 2 factors.
1! has 1 factor.

-----common sense--------

0! has no factors. But then again you can *1 as many times as you please... so 1* just...becomes 1 since the multiplication can't be performed because of absence of the second term?

I also heard an explanation that sounds like

"there is exactly one way to arrange zero objects"

which makes sense to me...I don't know. does this question have a solid, undebatable answer? or "it just is"?
 
Last edited:
  • #22
Robokapp said:
0! has no factors. But then again you can *1 as many times as you please... so 1* just...becomes 1 since the multiplication can't be performed because of absence of the second term?

I also heard an explanation that sounds like

"there is exactly one way to arrange zero objects"

Of course, an expty product is 1 by definition:

\prod_{\{\}}f=1
 
  • #23
from memory its something to do with the gamma distrubution
 
  • #24
FunkyDwarf said:
from memory its something to do with the gamma distrubution

The gamma function, not distribution.
 
  • #25
Firstly, we are free to define ! however we wish, and with the way we have done so n! naturally describes the number of ways of ordering n objects for n>0. The abstract function ! is *not* defined as the number of ways ordering objects. That might be why we found it useful in the first place, but that is not its definition.: things have a life beyond where we first find them in mathematics. If they didnt' they'd probably be forgotten.

Secondly there is no need to justifiy things by 'common sense'. Common sense is unique amongst virtues in that everyone possesses it (I forgot who said that). Mathematics is not common sense, in that sense. We are free to define whatever the heck we want, and this is the definition that makes most logical sense for us.
 
Back
Top