Do Residuals Always Sum to Zero in Regression Analysis?

  • Context: Graduate 
  • Thread starter Thread starter FallenApple
  • Start date Start date
  • Tags Tags
    Zero
Click For Summary

Discussion Overview

The discussion centers on whether residuals in regression analysis always sum to zero, particularly in the context of comparing two groups using different parameterizations of a regression model. The conversation explores the implications of model equivalence, estimation methods, and the conditions under which residuals may or may not sum to zero.

Discussion Character

  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • One participant suggests that in a regression model without an intercept, the residuals may not sum to zero, questioning the generality of the theorem regarding residuals summing to zero.
  • Another participant agrees that the two models presented are equivalent and posits that if the parameter estimates are equivalent, the residuals must also be identical, implying that if one model's residuals sum to zero, the other’s should as well.
  • Some participants assert that they are not aware of any theorem stating that residuals must sum to zero, noting that under Maximum Likelihood estimation, the sum of the products of residuals with certain regressors will equal zero instead.
  • There is a mention of the condition that if there is an intercept in the model, the residuals may sum to zero, but this is not universally accepted among participants.

Areas of Agreement / Disagreement

Participants express differing views on whether residuals must sum to zero, with some asserting that they do under certain conditions while others challenge this notion. The discussion remains unresolved regarding the existence of a theorem on this topic.

Contextual Notes

There are limitations regarding the assumptions made about model equivalence and the specific conditions under which residuals may sum to zero. The discussion also highlights the dependence on the estimation method used.

FallenApple
Messages
564
Reaction score
61
Ok so say I'm comparing two groups. I can do it this way ##Y_{i}=b_{1}*I(G1)+b_{2}*I(G2)+e_{i}## where I(G1) is 1 if in group 1 and 0 if not. I(G2) is 1 if in group 2 and 0 if not. In that case, my design matrix will not have a column of ones.

However, if I reparameterise to ##Y_{i}=b_{0}+b_{2}*I(G2)+e_{i}## since I know I(G2) and I(G1) has to sum to 1. I will get a design matrix with ones in the first column. I think there is a theorem that says that the residuals sum to 0 if this is the case.

Now, does this mean that the residuals sum to zero for the first parameterization as well? After all, the two models should be equivalent.
 
Physics news on Phys.org
The two models are equivalent, since ##G1=1-G2##, ##b_0=b_1## and ##b'_2=b_2-b_1## where ##b'_2## is the coefficient of ##I(G_2)## in the second model.

Given the models are equivalent, I imagine that the parameter estimates will be equivalent. Conceivably that may differ by estimation method. The method for OLS is Maximum Likelihood and I'm pretty sure that would give identical estimates, but one would need to work through the equations for the estimates, substituting the equivalences in the preceding paragraph, to be sure.

If the parameter estimates are equivalent then the residuals will be identical since the linear estimators will be identical, so if the residuals sum to zero for the first model they will do that for the second as well.
 
  • Like
Likes   Reactions: FallenApple and FactChecker
If you are minimizing the sum-squared-errors for your parameter estimates, I don't think that the residuals have to sum to zero.
 
  • Like
Likes   Reactions: FallenApple
I agree with FactChecker. I am not aware of any theorem about residuals summing to zero. If one is using Maximum-Likelihood estimation of the coefficients (ie the usual, simplest way) then the sum of the products of residuals ##\varepsilon_i## with regressors ##I(G2_i)## will be zero, that is, ##\sum_i \varepsilon_i I(G2_i)=0##. Could that be the theorem you had in mind?
 
  • Like
Likes   Reactions: FallenApple
andrewkirk said:
I agree with FactChecker. I am not aware of any theorem about residuals summing to zero. If one is using Maximum-Likelihood estimation of the coefficients (ie the usual, simplest way) then the sum of the products of residuals ##\varepsilon_i## with regressors ##I(G2_i)## will be zero, that is, ##\sum_i \varepsilon_i I(G2_i)=0##. Could that be the theorem you had in mind?

I thought that if there is an intercept or that there could be a transformation to the intercept, the residuals sum to 0.
 

Similar threads

  • · Replies 4 ·
Replies
4
Views
1K
  • · Replies 2 ·
Replies
2
Views
1K
Replies
3
Views
2K
  • · Replies 0 ·
Replies
0
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 8 ·
Replies
8
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 15 ·
Replies
15
Views
2K