Skip to main content

Section 10.7 Test Yourself

Checkpoint 10.7.1.

In a linear regression model, what does the R-squared value (coefficient of determination) represent?
  • The proportion of variation in the dependent variable that is accounted for by the independent variable(s).
  • Correct! The text explains that an R-squared of .8847 means the independent variable accounts for 88.47% of the variation in the dependent variable.
  • The slope of the line of best fit.
  • No, the slope is a separate value (the "Estimate" or "B-weight") that shows the direction and magnitude of the relationship.
  • The number of data points that are outliers.
  • No, R-squared is a measure of the overall fit of the model, not a count of specific points.
  • The probability that the model is correct.
  • No, R-squared measures how well the model explains the variation in the data, not the probability of it being true.

Checkpoint 10.7.2.

According to the chapter, what is a good rule of thumb for deciding if an independent variable is making an important contribution to the model’s prediction?
  • Check if the absolute value of its "t value" is larger than about 2.
  • Right! The text explains that the t-value helps determine if a predictor is contributing, using an absolute value greater than 2 as a benchmark.
  • Check if its "Estimate" (slope) is a positive number.
  • No, the sign of the slope (positive or negative) indicates the direction of the relationship, not its importance or statistical significance.
  • Check if the model’s R-squared value is greater than 0.90.
  • No, a high R-squared value indicates a good overall model fit, but it doesn’t tell you about the contribution of each individual variable.
  • Check if the variable is listed first in the summary output.
  • No, the order of variables in the summary output does not indicate their importance.

Checkpoint 10.7.3.

Under what condition is it appropriate to add `-1` to the model formula to force the line of best fit through the origin (0,0)?
  • When it is logical to assume that a value of zero for the independent variable implies a value of zero for the dependent variable.
  • Correct! The chapter uses the example that if teams have zero members, it’s logical to assume they will have zero attendance at a game.
  • It should be done in most models to make them simpler.
  • No, the chapter specifically warns that in most models, especially with survey data, this would not be appropriate.
  • Whenever you are analyzing data with a large number of outliers.
  • No, the presence of outliers does not determine whether forcing the intercept to zero is appropriate.
  • Only when there is just one independent variable in the model.
  • No, the logic depends on the nature of the variables, not the number of them.
Chapter Challenge
We intentionally ignored some of the output of these regression models, for the sake of simplicity. It would be quite valuable for you to understand those missing parts, however. In particular, we ignored the "p-values" associated with the t-tests on the slope/Bweight estimates and we also ignored the overall F-statistic reported at the very bottom of the output. There are tons of great resources on the web for explaining what these are.
For a super bonus, you could also investigate the meaning of the "Adjusted" r-squared that appears in the output.
You have attempted of activities on this page.