Prediction: fitting a predictive model to an observed dataset, then using that model to make predictions about an outcome from a new set of explanatory variables
Explanation: fit a model to explain the relationships between a set of variables
A foundational method for fitting a regression line<|>Principle: Minimize the sum of squared differences between observed values and those predicted by the line
y = β0 + β1 X + ϵ<|>Y is the outcome variable<|>X is the predictor variable<|>β0 is the intercept (the value of Y when X = 0)<|>β1 is the slope (the change in Y for a one-unit increase in X)<|>ϵ is the error term (the difference between the observed and predicted values of Y)
Represents the proportion of variance in the dependent variable (Y) explained by the independent variable (X)<|>Ranges from 0 to 1<|>Higher R-squared values indicate a stronger relationship between the variables
Takes into account the number of predictors in the model<|>Adjusts for the inclusion of additional predictors that may not improve the model's explanatory power<|>More useful in multiple regression, but can also be used in simple regression for comparison purposes
Extends simple linear regression to include multiple predictors, allowing for a more comprehensive understanding of relationships and improving prediction accuracy
β0 (Intercept): Expected value of y when all predictors are zero<|>β1, β2, . . . , βn (Coefficients): Expected change in y for a one-unit increase in the corresponding predictor, holding all other predictors constant
Enhanced understanding of the relationships between predictors and the dependent variable<|>Identification of potential confounding factors<|>Systematic approach to adding control variables in the model<|>Improved model interpretability