Learning theory - statistical learning
Today’s post is the first in a set of three that will discuss two theories of learning. In particular, today’s post will cover (a very narrow aspect of) statistical learning theory. The next post will present a game-theoretic view of learning, using online learning as its primary analytical framework. The final post will connect the two perspectives.
A bit of background
Learning problems, specifically machine learning problems can be analyzed from at least two perspectives. In either case, our goal is to have a toolbox from which we can analyze how well a learning system works. For example, we want to be able to say when learning occurs, when it doesn’t, and how quickly it occurs or doesn’t. To make this problem tractable, we’ll require
- a model for the data;
- assumptions on the source of that data;
- functions to generate predictions;
- loss functions to assess the quality of our predictions; and
- learning algorithms to help us improve our predictions by understanding our losses.
The set-up
We assume that our data consists of feature-output pairs with and . The data comes from some fixed but unknown source that could be stochastic (as in the statistical learning framework) or nonstochastic (as in online learning framework).
We a function that maps data to estimates/forecasts/predictions. So, gives us a mechanism to generate our own ’s that we can compare against the true ’s; we assess our estimates via a loss function . You probably have many loss functions that you know and love…
- squared-error loss: .
- absolute-error: .
- Kullback–Leibler: .
- hinge loss: .
- logistic loss: .
The final ingredient is a learning algorithm, which we can think of as a map from data to models. This is way that V. Vapnik and A. Chervonenkis conceptualized learning algorithms and we think that it’s a useful mental model.
Before diving into either framework let’s give ourselves a learning goal: our algorithms should output good models with respect to the data source.
Statistical learning theory
In the statistical learning setting, we assume that our data is generated by a probability (i.e., stochastic) model with distribution . In particular, we’ll assume
We are interested in learning by controlling the statistical risk
where the expectation is over the randomness in the data and the model belongs to some class of models . The class of models is a design parameter that we specify. For example, we could choose a set of linear predictors
Since we’re restricting our models to the class , we’re really interested in controlling our risk over that class. So we want a model
But this is statistics, so we don’t have access to at the population level. Instead, we have knowledge of through samples . Hmmm…
Empirical risk minimization
As statistically-minded individuals, we’ll use the empirical risk as a surrogate for the population risk and pick our model as
We want the empirical risk to a be a good proxy for the statistical risk. More specifically, we want it to be uniformly close to the statistical risk. By uniformly close, we mean that we’re close over all models within our class and all data.
This might seem a bit extreme, but it prevents us from being lulled into believing we have a “good” model by getting easy training data or having an overly complex model class—think interpolating the data.
Rademacher complexity
To help us assess our algorithms and models, we turn to Rademacher averages, which is a Vapnik–Chervonenkis type notion of complexity that (also) has its origins in the Russian school of statistics (~1970s). We defined the empirical Rademacher average
for a class of functions , samples . Before moving on, note that the empirical Rademacher average does not depend on , but only on the data. This will prove useful in the sequel.
Aside: see the Rademachers are rad - part I and Rademachers are rad - part II post for Rademacher fun in the spirit of Rademacher complexity.
Assessing empirical risk
Okay, now we’re going ready to assess how well empirical risk minimization works by studying the difference
which is the difference between a suboptimal model and the optimal model—note that we don’t have an empirical risk term… yet.
Nonetheless, here goes. We can add and subtract the empirical risk getting
The term is a comparison of the selected model on the whole distribution versus the sample. The second term is the error of using the “best” model on the training sample versus the whole distribution. The inequality follows from the and the fact that by definition of . Also, we can interpret the first term as a measurement of how much the model overfits the data. And, in some sense, this is what we want to understand!
High probability bounds
Both terms on the righthand side are random variables because they depend on the training data. So via Chernoff bounds (which we’re excluding from the current discussion), we can say that with probability
The expected value of the second term is 0 because the expected value of the empirical risk, by our i.i.d. assumption, is the statistical risk. You can think of this term as a bias-like term. The first term is more interesting; it may have caused your Rademacher neurons to fire, but it’s not quite a Rademacher average… yet.
The ghost sample
Now, we’re going to use one of “modern” statistics favorite tricks: the ghost sample. The ghost sample allows us to replace a (population) quantity by the sample quantity by pretending that we have access to a second, made-up sample from the distribution—it’s like we have a second training set. By the ghost sample trick (first equality) and Jensen’s inequality, our term is bounded as
Okay, cool. The righthand side is the difference between two empirical risk quantities, one for the original training sample and the other for the ghost sample. We’re getting closer to bounding the overfitting term…
Rademacher magic
Now, we introduce Rademacher random variables into the problem which will allow us to bound the overfitting. Think of the Rademachers as randomly permuting an example from the training set with an example from the ghost sample. We want to control how much the risk changes when we exchange examples and do so via the empirical Rademacher average
where and is the loss function for evaluated on example . Now let’s manipulate this quantity, first distributing terms and then exchanging the supremum with a sum to get an inequality:
So, this last calculation, the fact that has the same distribution as , and , imply that
Next up
… a short-and-sweet example of the material from this post and then the online learning perspective!
References
This post is related to material from Nicolò Cesa-Bianchi’s lectures at the Bordeaux Machine Learning Summer School in 2011. Watch them here: MLSS 2011 lectures.