Observing interactions with plots using est. coeff.

  • Context: Graduate 
  • Thread starter Thread starter FallenApple
  • Start date Start date
  • Tags Tags
    Interactions Plots
Click For Summary

Discussion Overview

The discussion revolves around the use of estimated coefficients in linear and logistic regression to assess interactions between groups. Participants explore the necessity of visualizing data alongside statistical tests, particularly focusing on how interaction is interpreted in both linear and logistic contexts.

Discussion Character

  • Debate/contested
  • Technical explanation
  • Mathematical reasoning

Main Points Raised

  • Some participants suggest that plotting regression lines using estimated coefficients can help visualize interactions between groups, while others question the necessity of visualization if p-values are available.
  • There is a discussion about the interpretation of p-values, particularly regarding which coefficients are being referenced and how p-values are calculated.
  • One participant argues against running separate models for each interaction category, proposing instead to find the p-value for the difference between models, emphasizing the importance of standard errors.
  • Concerns are raised about the validity of Wald's test, with a participant noting it may only be valid for models that are subgroups of each other.
  • Participants express skepticism about relying solely on p-values, suggesting that visualization of data can provide a clearer understanding, especially with large sample sizes where statistical tests may reject the null hypothesis too readily.
  • There is a proposal for using a likelihood ratio test between models as an alternative approach to assess interactions.
  • Questions are raised about the advantages of visualizing data over statistical tests, with some suggesting that perfect visualization could potentially eliminate the need for regression analysis.

Areas of Agreement / Disagreement

Participants do not reach a consensus on the best approach to assess interactions between groups. There are competing views on the importance of visualization versus reliance on p-values, and the validity of different statistical tests is debated.

Contextual Notes

Limitations in the discussion include the dependence on sample size and the sensitivity of statistical tests, as well as the unresolved nature of how to best interpret p-values in the context of interaction effects.

FallenApple
Messages
564
Reaction score
61
This question has two parts. On for the linear case, and one for the logistic case. Say X is a continuous variable and we want to see how x affects the response when looking between two different groups. Say G1=Group1, G2=Group2

In linear regression, we can plot the regression lines using the estimated coefficients to see if there is an interaction between two different groups. If they are parallel, then that suggests interaction, if they are not, then that suggests the opposite. Then I would check the p value of the coefficient to see if this is really the case.

Is that true? If it is, then why even plot using the estimated coefficients? The p values should be enough. if p!= 0 for the wald test for the interaction term, then there is insufficient evidence for interaction. Is it because if p=0, we still want to see just how much interaction there is? But wouldn't the absolute value of the interaction coeff be a good hint. Or do we still need visualization?What about for logistic regression. So I look at the probability curve, P[Y=1|X,G1] and P[Y=1|X,G2]. If the difference . delta =P[Y=1|X,G1] - P[Y=1|X,G2] is a constant at each X, then does that mean there is no interaction? Is this like the linear case?
 
Physics news on Phys.org
FallenApple said:
Then I would check the p value of the coefficient to see if this is really the case.

Which coefficient are you talking about? - and how do you arrive at a p-value for it?
 
It sounds like what the individual is doing is running p models for each category of interaction that the model may have. Then comparing the results of the model by their coefficients, and then derive a conclusion via the p-values. This is not the right approach. You need to instead find he p-value for the difference between the models. There's a standard error estimated between each model type, and it's very possible to get differences at each observed point but for the differences to not be statistically significant.

Wald's test would only be valid if the models are supgroups of each other. (Although don't quote me on that.)

Lastly relying on just p-values is never a good idea. If you can visualize your data, then do it. When your sample is large, nearly everything rejects the null hypothesis.
 
MarneMath said:
It sounds like what the individual is doing is running p models for each category of interaction that the model may have. Then comparing the results of the model by their coefficients, and then derive a conclusion via the p-values. This is not the right approach. You need to instead find he p-value for the difference between the models. There's a standard error estimated between each model type, and it's very possible to get differences at each observed point but for the differences to not be statistically significant.

Wald's test would only be valid if the models are supgroups of each other. (Although don't quote me on that.)

Lastly relying on just p-values is never a good idea. If you can visualize your data, then do it. When your sample is large, nearly everything rejects the null hypothesis.

Ok so basically do a likelihood ratio test between the two models right?

Also, why is visualizing data better than just getting p values from regression models? Is it because visualization looks at the data as it is? So if there is a way to perfectly visualize the data, when we would not need to do the regression at all?
 
As I stated, as your sample size increases, then most statistical test will reject the null hypothesis. Statistical test were designed to be rather sensitive. They weren't meant for millions upon millions of data points. Thus often times, for example, you'll reject Shapiro test, but if you look at the data, it's normal enough. You can even take smaller sub such that 99 times the Shapiro test fails to reject, but if you take the entire sample, it rejects.

Therefore, if possible, it's always good to look at your data and not rely on just statistical test.
 

Similar threads

Replies
3
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K
Replies
3
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 30 ·
2
Replies
30
Views
5K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 9 ·
Replies
9
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K