Every statistical software procedure that dummy codes predictor variables uses a default for choosing the reference category.

This default is usually the category that comes first or last alphabetically.

That may or may not be the best category to use, but fortunately you’re not stuck with the defaults.

So if you do choose, which one *should* you choose?

The first thing to remember is that ultimately, it doesn’t *really* matter, as long as you are aware of which category is the reference. You’re going to get the same results no matter what you choose. It’s just that the specific comparisons that the software reports (and gives you p-values for) will differ.

So it’s best to choose a category that makes interpretation of results easier. Here are a few common options for choosing a category.

Remember, the regression coefficients will give you the difference in means (and/or slopes if you’ve included an interaction term) between each other category and the reference category.

### Strategy 1: Use the normative category

In many cases, the most logical or important comparisons are to the most normative group. For example, in one data set I analyzed, an important dummy-coded predictor is Poverty Status: In Poverty or Not In Poverty.

Not In Poverty is the norm–most people aren’t in Poverty (at least in this data set–it may not be true in the population you’re studying). The interesting comparison is to see how people in poverty differ from this normative group. So making Not In Poverty the reference group just makes sense.

Likewise, another example is Marital Status: Never Married, Currently Married, Divorced, Separated, or Widowed.

The alphabetical default would make Widowed the reference group. But it’s not as interesting to compare Separated people to Widowed people, as they’re both small groups in the data set, and the most interesting comparisons are with the normative categories of Never Married or Currently Married.

In experiments or randomized control trials the control group is a natural normative category. The only exception I can think of is a study with multiple controls, but only one intervention or treatment group. In that case, it may be more important to measure any differences between the treatment and each control.

### Strategy 2: Use the largest category

The other problem with using the Widowed group as the reference is it’s very, very small. When sample sizes are very unequal in the groups, which is very common for naturally occurring groups, it can become problematic to use it as the reference.

Sometimes, if there isn’t a normative group in a logical sense, it makes sense to just use the largest category as the reference.

### Strategy 3: Use the category whose mean is in the middle, or conversely, at one of the ends

Sometimes all of these options fail. There is no obvious norm and sample sizes are similar.

In those cases, sometimes the best thing to do is to pick the category with the lowest, the highest, or the middle mean. Let me give you an example.

Let’s say those 5 marital categories have means on Y of

10 Never Married

11 Currently Married

9 Divorced

15 Separated

19 Widowed

If the overall F test in the ANOVA table is significant for this variable, you already know that the highest and lowest means are significantly different. You just don’t know which of the middle three are significantly different from each of those.

For example, the middle value here is 11, the mean for currently married folks. If you use that as the reference group and discover that it is significantly lower than 15, the mean for separated folks and 19, the mean for widowed, you know that both 9 for Divorced and 10 for Never Married should be too. (Note, this doesn’t always hold if some groups have much smaller sample sizes, but as long as they’re reasonably equal, it should hold).

You won’t know, for example, if there is a significant difference between the means for the Separated and Widowed groups, but if that’s not a theoretically important comparison, you’re done.

This particular strategy doesn’t always work, but you can use it to your advantage when it does.

Afreen says

I am confusing about reference category. why we change it. or it will be beneficial or not.

Karen Grace-Martin says

You don’t have to change it. Depending on your research question, sometimes changing it give you a comparison that most closely answers the question.

Afreen says

I have education variable . in education there are six category (no education, incomplete primary education, complete primary education, incomplete secondary education, complete secondary education and higher education) . If I am doing Poisson regression on stata. They automatically choose reference category as a no education. I want to check the effect of higher education will decrease the fertility. if I choose reference category is higher education then it will be good or not?

Karen Grace-Martin says

If you want to see if higher education has a lower mean fertility that the others, then yes, that would be a good idea.

Stijn says

Hi, I am observing data from june – mrt, now I want to see the effects of being in a specific month so i added dummies to the months (1, 0). As far as I now I should leave one month out of the regression which then becomes the base level? However if I do this and for example pick nov as a base (leave it out) then the results are highly significant. If I leave june out, the coefficients are the same but the results are insignificant for some variables. How is this possible?

Karen Grace-Martin says

I’m not sure if you mean the coefficients that have higher p-values for June are those for the months themselves or for other variables in the model. I’m assuming the latter. It could be a few different things. If June has a smaller sample size, using it as the base category can leave you with smaller power. Or, if the mean for June is right in the middle of the means, but November is the highest or lowest, there could be months whose means are different from Nov but not June.

Farah Mneimneh says

The coefficients of the control/dummy variables are not changing when I change the reference level of the independent variable. Is this normal? You mentioned that the coefficients might be affected if the reference level is of small counts compared to other levels.

Karen Grace-Martin says

I’m not sure if you mean the coefficient of the variable you’re changing or if you mean coefficients of other variables. If the one you’re changing reference levels for has only two values, its coefficient should change sign. Other coefficients shouldn’t change as long as there are no interactions.

GIRMA HUKA says

THANKS FOR CLEAR EXPLANATION!

GIRMA HUKA says

THANKS FOR CLEAR EXPLANATION.GIRMA HUKA DUKALE WEST GUJI ZONE,OROMIA,ETHIOPIA

Reema says

I am a beginner in data science. Just had a general doubt regarding the reference category.

So my doubt is whether the reference categories are always assumed to be significant by default as while giving business recommendatins we compare the remaining categories with reference eg so and so item “x” is more popular then the “reference category item” and hence client should consider producing more “x” items then the “reference item”.

Karen Grace-Martin says

Hi Reema,

It all depends on what you’re trying to test. If you’re trying to understand, say, whether customers like item A more than the reference item B, then yes, your test is about the difference in the mean of liking between those two items. The coefficient you care about then is item A’s. That’s the difference between A and B.

abdulaziz says

how can I select reference category in stata 9

Leonardo Castilho says

Why using small sample groups as reference is problematic?

Karen Grace-Martin says

Hi Leonardo,

It’s generally a lack of power. Your power is determined by your smaller group.

Rousset says

How do I chose the Reference Category in STATA, so that it is not arbitrary the last alphabetical one?

Presently, I am doing an xtreg in STATA and the omitted variable is the last one. I would like to chose another one so that results are easier to interpretate.

Thanks

Azadeh says

use ib#.[variable_name]. b stands for base and # is the number for that category in your variable.

Shalaw says

I am going to analyze a situation where there are 300 non-injury and only 17 injury… four categorical variables are significant according to Chi-squire, then I used Multiple logistic regression for significant variables. Three of them are significant again. does it make any sense? I would like to know whether can I use Multiple logistic regression because only 17 respondent had injured from 317 of the respondents. I used SPSS to data analysis.