Shukie said:
Thank you, that really cleared things up. I don't know how to calculate it in Mathematica, but I can do it with my graphic calculator.
normalcdf(0.0129, 10^9, 0, 0.00087) = 5*10^-50
I don't know if I calculated it correctly, because that would mean a mean of 0.0129 or higher would have a ridiculously low chance of happening, pretty much zero chance really. I don't know how realistic that is.
That sounds about right. Just check that the calculator not actually overflowing.
You've calculated sdev for b_i as 0.00292541
The standard deviation for 12 items, therefore, should be 0.000844493
(You've used 0.00087, which was a really crude low accuracy value by me, starting from 0.003)
You've got the mean as 0.0190084
That's about 22.5 standard deviations! 10^-50 sounds the right ball park.
In other words, there's a ridiculously low chance of no warming trend. You've definitely got warming happening; there is a statistically significant increase in temperature going on.
Now it just so happens that I play around with this kind of data quite a lot. I've loaded your De Bilt data into a spreadsheet I have ready to hand, and used that to calculate the trend with a linear regression over the data. But I used another slightly more complicated method for getting the significance of the trend.
I gave each year a single temperature as the mean of all twelve months. (This is already in your data as column 13). I did a regression over that, and found the slope. Then I used the "standard error" method, with the student-t distribution to get bounds from a confidence level. Here's what I got:
Warming trend is De Bilt: 108.000 year trend = 0.129 C/decade, +/- 0.039 (95% conf)
(My spreadsheet calculates trend in degrees per decade. You can divide by 10 to get degrees per year.)
The +/0 0.39 means that I am 95% confident the trend is somewhere from 0.090 to 0.168. There's too much noise to nail down an accurate slope better than that; but it's definitely positive.
The confidence limit is a parameter, and the spreadsheet cannot actually handle the extreme probability that would allow for a negative trend to be consistent with this data.
Here's how it works.
De Bilt: 108.000 year trend = 0.129 C/decade, +/- 0.051 (99% conf)
De Bilt: 108.000 year trend = 0.129 C/decade, +/- 0.066 (99.9% conf)
De Bilt: 108.000 year trend = 0.129 C/decade, +/- 0.079 (99.99% conf)
De Bilt: 108.000 year trend = 0.129 C/decade, +/- 0.091 (99.999% conf)
De Bilt: 108.000 year trend = 0.129 C/decade, +/- 0.102 (99.9999% conf)
De Bilt: 108.000 year trend = 0.129 C/decade, +/- 97957.433 (99.99999% conf) {Oops. Overflow}
The method you are using, with the mean and standard distribution for individual months, is a bit quicker and easier, and it does establish that the warming trend is unambiguous.
Cheers -- sylas
PS. You are in the Netherlands, I guess? You speak English like a native. Very impressive! I've been to Eindhoven and Amsterdam, back in 1997.
Added in edit. Thinking about this, I may have given you some bad advice. All the different months are not really "independent" samples, and so dividing by square root of 12 might have been inappropriate. I'm not sure; I am not a statistician. You could ask your instructor, or perhaps another reader here might correct me on this.
Suppose you just stick with 0.00292541 as the standard deviation, rather than 0.00087 or whatever. The method I used ends up with 95% confidence limits at 0.0039; that suggests to me that dividing by sqrt(12) might have given just a bit TOO much confidence in the warming.
In this case, you would find that the mean is about 6.5 standard deviations above zero. This is still enough to give very high confidence for warming, but it might be the more sensible bound, statistically. I'm really not sure! Anyhow, plug that into your normalcdf as well.