Solution Concepts

Linear Equations > Definitions | Properties | Solution concepts
  • Solution set

  • Least-norm solutions

  • Approximate solutions

Solution set

We consider the linear equation

 Ax = y,

where A in mathbf{R}^{m times n} and y in mathbf{R}^m are given. The associated solution set is the set of solutions

 mathbf{S} := left{ x ~:~ Ax = y right}.

By construction, it is an affine set of mathbf{R}^m.

Several cases may occur:

  • If y is in the range of A, the equation has at least a solution, and the solution set is not empty. If the nullspace is not {0}, then this solution is not unique. In this case, we develop the notion of minimum-norm solution, which allows to select a particular solution among the many.

  • If y is not in the range of A, then there are no solutions: the solution set is empty. In this case we introduce the notion of approximate solution, which is based on minimizing the norm of the ‘‘residual error’’ Ax-y. We refer to a vector x that minimizes this error in norm, as a minimum-residual solution (although, strictly speaking, it is not a solution to the original equation).

For both the minimum-norm and minimum-residual problems, the qualitative behavior of the solution depends crucially on the norm used to measure the size of the vectors involved.

Minimum-norm solutions

In the case when y in mathbf{R}(A), there may be many solutions to the linear equation Ax = y, namely if and only if the nullspace of A is not reduced to {0}. It is then of interest to select a specific solution. A popular choice is the minimum-norm solution, obtained via the optimization problem

 min_x : |x| ~:~ Ax = y.

Again, the qualitative behavior of the solution depends heavily on the choice of the norm used. Depending on the choice for the norm |cdot|, such problems may or may not be easy to solve. For the popular norms (l_p, with p in {1,2,infty}), they are.

A popular choice corresponds to the case when the norm in the objective function of the above problem is the Euclidean norm. When A is full row rank, meaning that there is always a solution irrespective of y, the minimum-Euclidean-norm solution has a closed-form expression

 x_{rm MN} = A^T(AA^T)^{-1} y .

(Note that the inverse exist, since A is full row rank.) The geometric interpretation of this solution is that it is the projection of 0 on the set of solutions to the linear equation.

Example: Control of a unit mass.

Other norms are often used. For example, the l_1 norm is often used as it tends to produce sparse solution vectors x, meaning a vector x which has many zero components. This is in turn useful when one is trying to achieve some desired target y=Ax by proper choice of x, with the smallest number of non-zero components.

Example: Sparse image compression.

Approximate solutions via residual error minimization

Basic idea

When y notin mathbf{R}(A), there exist no solutions to the equation Ax = y. This can be due to several factors: the linear equation may be an approximation of a non-linear one, resulting in model errors; or y might be a noisy measurement of the output of some linear process, resulting in measurement errors; or both. Such errors might make the solution set empty.

We can then try to find an approximate solution, by minimizing the distance from Ax to y by proper choice of x. We are led to problems of the form

 min_x : |Ax - y|.

Again, the ‘‘solution’’ found this way depends heavily on the choice of norm. Also that choice has an impact on the computational complexity of the optimization problem above. As before for the popular norms (l_p, with p in {1,2,infty}), these problems are easy.

Least-squares

The most well-known case corresponds to the choice of the l_2 (Euclidean) norm, because it is the only one of the three which leads to a closed-form solution. The corresponding problem

 min_x : |Ax-y|_2^2.

is called a least-squares problem. We will see later that, when A is full column rank, the problem has a unique, closed-form solution, which is in the form

 x_{rm LS} = (A^TA)^{-1}A^T y.

Examples:

Other norms

Other norms are often used. For example one can experimentally observe that using the l_1 norm tends to produce solutions such that (Ax)_i = y_i for many i's: the residual error vector Ax-y is sparse (has many zero elements). Solving the corresponding problem can be efficiently done, but no ‘‘closed-form’’ solution is known, so we must use optimization solvers. We examine the corresponding problem here.