In this post, we’re going to be Bayesians—well, sort of… actually we’re going to be empirical Bayesians. If you’ve never seen empirical Bayes, then get ready for a beautiful way to blend the best of frequentist and Bayesian style thinking. And by the way, it’s super useful in many “modern” problems.

Before going any further down Empirical Bayes Avenue, it’s worth saying that empirical Bayes is not really a Fisherian idea. But Fisher wasn’t always Fisherian, so I think we’re justified in discussing an idea that is not attributed to Fisher; moreover it is spiritually similar, in some sense, to Fisher’s style—more on that in the next post.

The key idea of empirical Bayes is to share statistical strength, which is a very natural thing for Bayesian type analysis, but to do so while respecting the underlying data—this is the empirical part. By respecting the underlying data, I mean that we will use the underlying to estimate our prior distribution.

Teeing things up

We’re going to quickly review a few Bayesian ideas before jumping in to empirical Bayes. Assume we have a parameter drawn from a prior distribution . We observe data that is generated by a density . Notationally, we have

After having observed the data , we estimate the posterior distribution of via Bayes rule

To be empirical Bayesian, we perform the sleight of hand

and replace our prior with an estimate of it that is based on the observed data . Without getting too tied up in details about how we estimate , let’s look at an example taken from Efron’s paper.

Efron’s log-odds example

In the prior post, we wanted to estimate the sample log-odds. We calculated it as

These data are from a hospital that was testing a new treatment and similar data exist for 40 other hospitals. Using the other data, we can estimate confidence intervals for each hospitals log-odds estimate. Doing so gives a 90% (central a posteriori) confidence interval of

The ordinary, frequentist-style 90% interval is

This interval is estimated using only the data in hospital 8’s table. Notice that the ordinary interval is shifted to the left of the empirical Bayesian interval. This results from the fact that majority of the other data produces log-odds estimates which are greater than -4.2. Indeed, the only estimates smaller than this are .

Next up

… a recap on Fisher’s style and contributions!

References

This post is related to material from:

  • “R.A. Fisher in the 21st Century” by Bradley Efron.
  • Computer Age Statistical Inference: Algorithms, Evidence, and Data Science by Bradley Efron and Trevor Hastie. A digital copy lives here: CASI.
  • Large Scale Inference by Bradley Efron.