## Supporting documents

Here are the slides of the presentation. The videos are on eCampus.

## Abstract

Two central problems in machine learning are

1. Data fitting (learning a function),
2. Data classification.

### Data fitting

Learning a function is the following problem: given a list ((x1, y1), …, (xm, ym)), xiRn, yiR, viewed as input/output of some unknown function, find a “nice” function f: RnR such that f(xi) = yi. If we look for an affine function f(x) = wT x + b, then we can view the problem as an optimization problem: minimize the error ∑i=1m (yixiT wb)2.

The difficulty is that since w is unconstrained, this minimization problem “blows up.” Thus it is necessary is to add some regularization term. We will discuss three scenarios:

• (i) Add an l2 term K wT w. This is ridge regression.
• (ii) Add an l1 term τ ∑i=1m |wi|. This is lasso.
• (iii) Add both an l1 term and an l2 term. This is elastic net.

Problems (ii) and (iii) do not have a closed form solution. They can be solved using a powerful iterative method, ADMM.

### Data classification

The problem of data classification is that we have two disjoint sets of data points, {u1, …, up} (the blue points), and {v1, …, vq} (the red points), and we want to find a hyperplane H of equation wT xb = 0 such that the blue points and the red points belong to the two disjoint open half spaces determined by H.

If this is possible, the method SVM due to Vapnik picks H as the hyperplane that maximizes the least distance of the blue and the red points to H. The problem can be formulated as an optimization problem that can be solved using its Lagrangian dual.

When separation is impossible, we relax the separation criterion by allowing a soft margin of error (points can invade the wrong side of the hyperplane). Again, this problem (soft marging SVM) can be formulated as an optimization problem and solved using its dual (using ADMM).

We will sketch how the Lagrangian framework dealing with inequality constraints is used to solve the SVM (hard and soft) problems.