Undergrad Vector visualization of multicollinearity

  • Thread starter Thread starter Trollfaz
  • Start date Start date
  • Tags Tags
    Regression analysis
Click For Summary
In regression analysis, multicollinearity occurs when there is significant correlation among the predictor variables, which can complicate the estimation of coefficients. A general linear model is represented as y = a_0 + Σ a_i x_i, where n observations of y are collected for k predictors. Perfect multicollinearity arises when at least one predictor is a linear combination of others, resulting in a rank deficiency in the matrix X, which is structured with observations as rows and predictors as columns. Strong multicollinearity indicates that some predictors make small angles with the subspace formed by others, impacting the model's reliability. Understanding and addressing multicollinearity is crucial for accurate regression analysis and interpretation.
Trollfaz
Messages
143
Reaction score
14
General linear model is
$$y=a_0+\sum_{i=1}^{i=k} a_i x_i$$
In regression analysis one always collects n observations of y at different inputs of ##x_i##s. n>>k or there will be many problems. For each regressor, and response y ,we tabulate all observations in a vector ##\textbf{x}_i## and ##\textbf{y}_i##, both is a vector of ##R^n##.So multicollinearity is the problem that there's significant correlation between the ##x_i##s. In practice some degree of multicollinearity exists. So perfectly no multicollinearity means all the ##\textbf{x}_i## are orthogonal to each other?ie.
$$\textbf{x}_i•\textbf{x}_j=0$$
For different i,j and strong multicollinearity means one of more of the vector makes a very small angle with the subspace form by the other vectors? As far as I know perfect multicollinearity means rank(X)<k. X is a n by k matrix with ith col as ##\textbf{x}_i##
 
Physics news on Phys.org
Perfect multicollinarity means that at least 1 predictor variable (columns) is a perfect linear combination of one or more of the other variables. Typically the variables are the columns of the matrix and observations are rows. In this situation, the matrix will not be full rank.
 
  • Like
Likes FactChecker
The standard _A " operator" maps a Null Hypothesis Ho into a decision set { Do not reject:=1 and reject :=0}. In this sense ( HA)_A , makes no sense. Since H0, HA aren't exhaustive, can we find an alternative operator, _A' , so that ( H_A)_A' makes sense? Isn't Pearson Neyman related to this? Hope I'm making sense. Edit: I was motivated by a superficial similarity of the idea with double transposition of matrices M, with ## (M^{T})^{T}=M##, and just wanted to see if it made sense to talk...

Similar threads

  • · Replies 0 ·
Replies
0
Views
2K
  • · Replies 0 ·
Replies
0
Views
814
  • · Replies 25 ·
Replies
25
Views
3K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 12 ·
Replies
12
Views
3K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 4 ·
Replies
4
Views
4K