Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

What is the difference between Multiple R-squared and Adjusted R-squared in a single-variate least squares regression?

Could someone explain to the statistically naive what the difference between Multiple R-squared and Adjusted R-squared is? I am doing a single-variate regression analysis as follows:

 v.lm <- lm(epm ~ n_days, data=v)  print(summary(v.lm)) 

Results:

Call: lm(formula = epm ~ n_days, data = v)  Residuals:     Min      1Q  Median      3Q     Max  -693.59 -325.79   53.34  302.46  964.95   Coefficients:             Estimate Std. Error t value Pr(>|t|)     (Intercept)  2550.39      92.15  27.677   <2e-16 *** n_days        -13.12       5.39  -2.433   0.0216 *   --- Signif. codes:  0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1   Residual standard error: 410.1 on 28 degrees of freedom Multiple R-squared: 0.1746,     Adjusted R-squared: 0.1451  F-statistic: 5.921 on 1 and 28 DF,  p-value: 0.0216  
like image 862
fmark Avatar asked May 20 '10 02:05

fmark


People also ask

What is the difference between multiple R-squared and adjusted R-squared?

What Is the Difference Between R-Squared and Adjusted R-Squared? The most vital difference between adjusted R-squared and R-squared is simply that adjusted R-squared considers and tests different independent variables against the model and R-squared does not.

How does the difference between R square and the adjusted R square change as the sample size increases?

In general, as sample size increases, the difference between expected adjusted r-squared and expected r-squared approaches zero; in theory this is because expected r-squared becomes less biased.

Do you report multiple R-squared or adjusted R-squared?

The short answer is that you should almost always report adjusted R-squared in favor of R-squared.

When to use R-squared and adjusted R-squared?

Difference between R-square and Adjusted R-square Every time you add a independent variable to a model, the R-squared increases, even if the independent variable is insignificant. It never declines. Whereas Adjusted R-squared increases only when independent variable is significant and affects dependent variable.


1 Answers

The "adjustment" in adjusted R-squared is related to the number of variables and the number of observations.

If you keep adding variables (predictors) to your model, R-squared will improve - that is, the predictors will appear to explain the variance - but some of that improvement may be due to chance alone. So adjusted R-squared tries to correct for this, by taking into account the ratio (N-1)/(N-k-1) where N = number of observations and k = number of variables (predictors).

It's probably not a concern in your case, since you have a single variate.

Some references:

  1. How high, R-squared?
  2. Goodness of fit statistics
  3. Multiple regression
  4. Re: What is "Adjusted R^2" in Multiple Regression
like image 197
neilfws Avatar answered Sep 20 '22 11:09

neilfws