How Reliable Is Your Anti-Spam Software's Error Detection?

  • Thread starter Thread starter moonman239
  • Start date Start date
  • Tags Tags
    Errors Type
moonman239
Messages
276
Reaction score
0
Let's say I'm testing anti-spam software. The number of false positives (aka, friendly messages misidentified as spam, for those who don't know the term) is 40. The number of false negatives (spam messages misidentified as friendly) is also 40. I'm testing 100 messages. How many more messages would I need to test in order to be 99.99% that the null hypothesis can/cannot be rejected?
 
Last edited:
Physics news on Phys.org
Thanks for the link, but I'm not seeing any equation whatsoever that will help. As far as I can see, all the listed equations have to do with means.
 
Oh, so what's your null hypothesis? I thought you were to test that the average message is not spam.
 
Last edited:
I just don't see how a sample mean would be relevant. Your understanding of the question is correct.
 
Be very specific, please: What is your null hypothesis?

Also, you have told us about 80 messages, what about the other 20?
 
D H said:
Be very specific, please: What is your null hypothesis?

Also, you have told us about 80 messages, what about the other 20?

Let's say my null hypothesis is that a spam message will be correctly marked as spam. As for the 20 messages, let's say that those are false negatives.
 
You already said you had 40 false positives and 40 false negatives out of 100 tests. That makes for a total of 80 out of 100. Those remaining 20 are either true positives or true negatives.

What does the confusion matrix for your test results look like?
 

Similar threads

Back
Top