IPM Part 2: The Lee-Sidford Barrier

  • These notes are based on my semester research project in the Theoretical Computer Science Lab at EPFL, where I was supervised by Prof. Kapralov and Kshiteej Sheth.
  • To understand the context in which the objects presented here are used, read my notes on Interior Point Methods.

1. Warm-Up: The Logarithmic Barrier

One simple barrier that one can consider for linear constraints is the logarithmic barrier :

Definition 1: Logarithmic barrier - Let the -dimentional polytope with constraints . The logarithmic barrier is defined as

To study its self-concordance, we start by proving the following lemma

Lemma 1: If and are - and -self-concordant barriers on and respectively, then is a -self-concordant barrier on .

Proof. The two properties of -self-concordance (Definition 4) are validated by just combining the inequalities of self-concordance of and (and using triangle inequality for the second property).


Lemma 2: The Logarithmic barrier is -self-concordant.


Proof. Take any and . Let .

and we just proved the first property of self-concordance for .
Now,

Thus, is 1-self-concordant, for any .

Since and by Lemma 1., is -self-concordant.

From our IPM framework, we conclude that using the logarithmic barrier yields an iteration algorithm for solving Linear Programs. The main problem with this is that since can get exponential in , this can get very inefficient, for example when the LP has a lot of repeated (or very similar) constraints. The next logical step is to try and reweigh the constraints, and give less weight to the ones that are repeated.

2. The Weighted Logarithmic Barrier

We are looking for a barrier that behaves like for to be determined later. We thus define and we will study its self concordance.

Lemma 3: We have

And

where is the leverage score of , namely

Proof. We have that and . Hence,

But is an orthogonal projection matrix, so its operator norm is smaller than . We can then conclude that

For the second part of the proof, let be the th slack condition. We have

Furthermore,

Hence, we have

Rescaling by a constant, we get that

Which gives us a self-concordance factor of . It depends on , but if , note that we have that because of the properties of leverage scores.

Hence with this intuition, we can see how one can try to look for a self-concordant barrier. The next section will present the actual barrier found by Lee and Sidford, and give some details about it.

3. The Lee-Sidford Barrier

As seen in the previous section, we want to weight the th constraint with a weight such that . This corresponds exactly to the definition of the Lewis weights. By their recursive nature, it's very hard to compute them exactly, so one can try to relax those condition to taking to be the vector of the Lewis weights, with large enough.

Lee and Sidford introduced the following barrier :

And proved the following result :


Theorem 1: If has full rank, then for any , is a self concordant barrier. In particular, for is a self concordant barrier.


The proof is lengthy and can be found in Reference 1, but one insight that might be interesting is the following (proof omitted):

Lemma 4: We have

Where is the th Lewis weight of .

This lemma is insightful because it shows that the hessian of is spectrally close to , which is the Hessian of the Lewis weights reweighted logarithmic barrier, and this can justify why the Lee-Sidford barrier behaves as wanted.

4. Putting it all together

We have exhibited a self-concordant barrier, which can be used to get a iteration algorithm to solve Linear Programs. One important thing to study now is the per iteration cost, in particular the cost to compute the barrier. The Lee-Sidford is very costly to compute because it involves determinants, and Lewis weights. Not being careful about how one computes it might ruin all our previous efforts.

The idea is to compute the barrier iteratively: the broad picture is that from , , and , if and are close enough, one can estimate with satisfactory precision, because is smooth enough. Furthermore, computing from only takes a logarithmic number of linear systems to solve. We can now state the final result of Lee and Sidford :


Theorem 2: Given an interior point of a Linear Program, there exists an algorithm that outputs a feasible point such that with constant probability in time, where is the work needed to compute for a positive diagonal matrix and a vector .


Depending on the nature of the Linear Program, might vary, and as such, the most general result with state of the art techniques states that one can solve approximately a Linear Program in time where is the constant multiplication matrix.

References

  1. Y. T. Lee, A. Sidford - Solving Linear Programs with sqrt(rank) linear system solves
Last Updated: