Generalized Linear Models and Their Connection to Ordinary Least Squares Linear Regression

Author
Affiliation

Alex Kaizer

University of Colorado-Anschutz Medical Campus

This page is part of the University of Colorado-Anschutz Medical Campus’ BIOS 6618 Recitation collection. To view other questions, you can view the BIOS 6618 Recitation collection page or use the search bar to look for keywords.

Generalized Linear Models

It turns out much of what we have done this semester with linear regression can be directly related to a larger family of models known as generalized linear models. These include a family and link function.

Families refer to the type of outcome:

  • Gaussian (i.e., normal) for continuous outcomes
  • Gamma for continuous outcomes restricted to positive values (also often skewed)
  • Binomial for binary outcomes
  • Poisson for count outcomes
  • Quasi-models (binomial, Poisson) that relax the dispersion parameter of our exponential families (i.e., are a little more flexible)

The link function connects how we estimate our outcome from our predictors and can vary by family. For Gaussian models there are three options for underlying models:

  • Identity: \(Y = \beta_0 + \beta_1 X_1 + \beta_2 X_2\)
  • Inverse: \(\frac{1}{Y} = \beta_0 + \beta_1 X_1 + \beta_2 X_2 \implies Y = \frac{1}{\beta_0 + \beta_1 X_1 + \beta_2 X_2}\)
  • Log: \(\log(Y) = \beta_0 + \beta_1 X_1 + \beta_2 X_2 \implies Y = \exp(\beta_0 + \beta_1 X_1 + \beta_2 X_2) = \exp(\beta_0) \times \exp(\beta_1 X_1) \times \exp(\beta_2 X_2)\) (i.e., our multiplicative interpretation since the relationship between Y and \(\mathbf{X}\) is no longer additive)

Let’s look at some examples of these and how they sometimes connect to other topics from the semester.