One of the big assumptions of linear models is that the residuals are normally distributed. This doesn’t mean that Y, the response variable, has to also be normally distributed, but it does have to be continuous, unbounded and measured on an interval or ratio scale.

Unfortunately, categorical response variables are none of these.

No matter how many transformations you try, you’re just never going to get normal residuals from a model with a categorical response variable.

There are a number of alternatives though, and one of the most popular is logistic regression.

In many ways, logistic regression is very similar to linear regression. One big difference, though, is the logit link function.

### The Logit Link Function

A link function is simply a function of the mean of the response variable Y that we use as the response instead of Y itself.

All that means is when Y is categorical, we use the logit of Y as the response in our regression equation instead of just Y:

The logit function is the natural log of the odds that Y equals one of the categories. For mathematical simplicity, we’re going to assume Y has only two categories and code them as 0 and 1.

This is entirely arbitrary–we could have used any numbers. But these make the math work out nicely, so let’s stick with them.

P is defined as the probability that Y=1. So for example, those Xs could be specific risk factors, like age, high blood pressure, and cholesterol level, and P would be the probability that a patient develops heart disease.

### Why Bother With This Logit Function?

Well, if we used Y as the outcome variable and tried to fit a line, it wouldn’t be a very good representation of the relationship. The following graph shows an attempt to fit a line between one X variable and a binary outcome Y.

You can see a relationship there–higher values of X are associated with more 0s and lower values of X have more 1s. But it’s not a linear relationship.

Okay, fine. But why mess with logs and odds? Why not just use P as the outcome variable? Everyone understands probability.

Here’s the same graph with probability on the Y axis:

It’s closer to being linear, but it’s still not quite there. Instead of a linear relationship between X and P, we have a sigmoidal or S-shaped relationship.

But it turns out that there are a few functions of P that do form reasonably linear relationships with X. These include:

- Square root of arcsin
- Complimentary log-log
- Probit
- Logit

The logit function is particularly popular because, believe it or not, its results are relatively easy to interpret. But many of the others work just as well.

Once we fit this model, we can then back-transform the estimated regression coefficients off of a log scale so that we can interpret the conditional effects of each X.

{ 4 comments… read them below or add one }

What will be the form of the logit function in case that Y is a binary variable such that y=0 with probability p and y=1 with probability (1-p).

Thanks

Is it possible negative slope of the line in the chart probit؟

Hello,

I am some doubt regarding adding interaction term in a model. When we add interaction term in a model and how to interpret the coefficient of interaction. If you have some paper or book, kindly send it to me. It helps me a lot.

Thanks

Hi Raj,

I’m not sure I can think of anything written on interpreting interactions in logistic regression, but we do cover this in the logistic regression workshop.

The very basic idea, though, is that the odds ratio for an interaction is the ratio of odds ratios. It’s hard to explain without knowing if the terms you are interacting are continuous or categorical, but that’s the basic definition. It takes me a good half hour to go over this in the workshop. 🙂