In this post, we’re going to talk about a concept that is useful in many areas… statistics, optimization, life-in-general… I call it idea chaining. My guess is that most people are already aware of this notion on an intuitive level. Systems thinkers would probably call it something like “problem decomposition”, mathematicians would think of lemmas… basically we take a problem, split it into smaller, easier-to-solve sub-problems, and then chain our results together to solve the original problem.

A (convex) calculus for convexity verification

In convex analysis and optimization, we’re often interested in whether or not a given function is convex. There are a few ways to go about showing the “vexity” of a particular function, such as using:

  • the definition of convexity;
  • the restriction-to-a-line technique;
  • first-order (i.e., derivative) conditions;
  • second-order conditions; or
  • a convex calculus.

The convex calculus approach is appealing because it gives us a toolbox to (mindlessly) construct or decompose a function, using a set of basic functions and a few rules that preserve convexity. In the context of convex optimization, this constructive convex verification process is part of an approach known as disciplined convex programming (DCP). The DCP ideas arose from the work of Michael Grant and Stephen Boyd in the mid-2000s. Hiriart-Urruty and Lemaréchal discuss the notion of a convex calculus in their books from the 1990s.

Let’s put the ideas to work on an example that came up on the quiz page of the DCP website (which teaches the principles of disciplined convex programming).

Our function of interest is

where

At first sight, this looks ugly because we have a concave function with , a couple convex functions with and , an affine function , and that Huber function (which is actually convex, but may not be recognized as such). Moreover, we have four variables!

The DCP / convex calculus approach turns this problem into an easy verification process. The non-obvious and key ingredient is the “log-sum-exp” function, also called the soft-max function. For , the log-sum-exp function is

This function is convex—if you’re suspicious, calculate its Hessian and convince yourself. With this single fact, we’re nearly done because the functions appearing within the exponentials are convex and affine. Then we use that log-sum-exp is convex and we’re done.

Tips and tricks. Here are the DCP rules we used:

  • an increasing convex function of a convex function is convex; and
  • a sum of convex functions is convex.

So, with two rules and a few basic functions we’ve show a rather complicated function to be convex—no gradients, subgradients, or Hessians required. We reduced the problem to a few simple subproblems (e.g., verifying the convexity of and the like) and then strung everything together using a couple rules. Bam.

If you’re interested in seeing the “vexity” parse tree for this function, check out the DCP site’s “Analyzer” and type in log_sum_exp(exp(max(u, z)), huber(w) + y - 42). Do it!

Le Cam’s inequality

Our next example of idea chaining stems from a problem from Tom Ferguson’s (excellent) mathematical statistics book called A Course in Large Sample Theory. Problem 5 of chapter 3 (“Convergence in Law”) walks us through Le Cam’s inequality.

Here’s the set up.

  • We have independent Bernoulli random variables each with its own probability of success, . We’ll call these .
  • We define .
  • We define as a Poisson random variable with parameter .

We want to show that

for all sets . In other words, we’re computing a bound on the error of using a Poisson approximation.

We’ll solve this problem by solving three simpler sub-problems.

(1) First, we’ll show that

(2) Then, we’ll show that

(3) Finally, we’ll show that

To show (1), (2), and (3), we’re going to couple and to the same underlying probability space.

  • The variables are uniformly distributed on .
  • We can define with .
  • We introduce as

where is the cumulative distribution function for a Poisson random variable with parameter .

Proof of (1).

Without loss of generality we can assume that . This allows us to conclude that . Now, we isolate the set where by noting that

One down, two to go.

Proof of (2).

First, let’s rewrite as . This implies that there’s at least one not equal to . (The at least phrase should be triggering the part of your brain associated with unions.) So, we have that

by the subadditivity of probability.

Two down, one to go.

Proof of (3).

Using our definitions of and , we can write

Now, we sum both sides of the equation and we’ve got it.

Bringing it all back home.

Finally, chain the inequalities in (1), (2), and (3) together and we’ve proved Le Cam’s inequality. Heck yeah!

References

This post draws on material from:

  • Stephen Boyd’s “Convex Optimization Short Course” notes;
  • Fundamentals of Convex Analysis by Hiriart-Urruty and Lemaréchal; and
  • A Course in Large Sample Theory by Thomas Ferguson.