# Linear systems¶

We now attend to the central problem of this chapter: Given a square, \(n\times n\) matrix \(\mathbf{A}\) and an \(n\)-vector \(\mathbf{b}\), find an \(n\)-vector \(\mathbf{x}\) such that \(\mathbf{A}\mathbf{x}=\mathbf{b}\). Writing out these equations, we obtain

If \(\mathbf{A}\) is nonsingular, then the mathematical expression of the solution is \(\mathbf{x}=\mathbf{A}^{-1}\mathbf{b}\), because

When \(\mathbf{A}\) is singular, then \(\mathbf{A}\mathbf{x}=\mathbf{b}\) may have no solution or infinitely many solutions.

If we define

then it is easy to check that for any real value of \(\alpha\) we have

Hence the linear system \(\mathbf{S}\mathbf{x}=\mathbf{b}\) with \(\mathbf{b}=\begin{bmatrix} 1\\0\end{bmatrix}\) has infinitely many solutions. For many other choices of \(\mathbf{b}\) the system can be proven to have no solutions.

## Don’t use the inverse¶

Matrix inverses are indispensable for mathematical discussion and derivations. However, as you may remember from a linear algebra course, they are not trivial to compute from the entries of the original matrix. While it can be done numerically by computer, it almost never is, because when the goal is to solve a linear system of equations, the inverse is not needed—and the process of finding it is slower than solving the original problem.

Julia does have a command inv that finds the inverse of a matrix. But, as demonstrated in Interpolating the population of China, in order to solve a linear system of equations, you should use backslash (the `\`

symbol, not to be confused with the slash `/`

used in web addresses).

## Triangular systems¶

The solution process is especially easy to demonstrate for a system with a triangular matrix. For example, consider the lower triangular system

The first row of this system states simply that \(4x_1=8\), which is easily solved as \(x_1=8/4=2\). Now, the second row states that \(3x_1-x_2=5\). As \(x_1\) is already known, it can be replaced to find that \(x_2 = -(5-3\cdot 2)=1\). Similarly, the third row gives \(x_3=(0+1\cdot 2)/3 = 2/3\), and the last row yields \(x_4=(1-1\cdot 2 + 1\cdot 1 + 1\cdot 2/3)/2 = 1/3\). Hence the solution is

The process just described is called forward substitution. In the \(4\times 4\) lower triangular case of \(\mathbf{L}\mathbf{x}=\mathbf{b}\) it leads to the formulas

For upper triangular systems \(\mathbf{U}\mathbf{x}=\mathbf{b}\) an analogous process of backward substitution begins by solving for the last component \(x_n=b_n/U_{nn}\) and working backward. For the \(4\times 4\) case we have

Solving the system backward, starting with \(x_4\) first and then proceeding in descending order, gives

It should be clear that forward or backward substitution fails if, and only if, one of the diagonal entries of the system matrix is zero. We have essentially proved the following theorem.

A triangular matrix is singular if and only if at least one of its diagonal elements is zero.

## Implementation¶

Consider how to implement the sequential process implied by equation (27). It seems clear that we want to loop through the elements of \(\mathbf{x}\) in order. Within each iteration of that loop, we have an expression whose length depends on the iteration number. One way we could do this would be with a nested loop:

```
for i = 1:4
s = 0
for j = 1:i-1
s += L[i,j]*x[j]
end
x[i] = (b[i]-s) / L[i,i]
end
```

A briefer version of the inner loop over `j`

is the comprehension

```
s = sum( L[i,j]*x[j] for j=1:i-1 )
```

However, when `i`

equals 1, the range `1:i-1`

is empty and the sum fails. To avoid this we can handle this case before the `i`

loop begins, and start that loop at 2. This is the approach taken in Function 12.

**Forward substitution to solve a lower-triangular linear system**

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 | ```
"""
forwardsub(L,b)
Solve the lower-triangular linear system with matrix `L` and
right-hand side vector `b`.
"""
function forwardsub(L,b)
n = size(L,1)
x = zeros(n)
x[1] = b[1]/L[1,1]
for i = 2:n
s = sum( L[i,j]*x[j] for j=1:i-1 )
x[i] = ( b[i] - s ) / L[i,i]
end
return x
end
``` |

The implementation of backward substitution is much like forward substitution and is given in Function 13.

**Backward substitution to solve an upper-triangular linear system**

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 | ```
"""
backsub(U,b)
Solve the upper-triangular linear system with matrix `U` and
right-hand side vector `b`.
"""
function backsub(U,b)
n = size(U,1)
x = zeros(n)
x[n] = b[n]/U[n,n]
for i = n-1:-1:1
s = sum( U[i,j]*x[j] for j=i+1:n )
x[i] = ( b[i] - s ) / U[i,i]
end
return x
end
``` |

The example in Solving triangular systems is our first clue that linear system problems may have large condition numbers, making inaccurate solutions inevitable in floating point arithmetic. We will learn how to spot such problems later in the chapter. Before reaching that point, however, we need to discuss how to solve general linear systems, not just triangular ones.

## Exercises¶

✍ Find a right-hand side vector \(\mathbf{b}\) such that the system \(\begin{bmatrix} 0&1\\0&0 \end{bmatrix} \mathbf{x}=\mathbf{b}\) has no solution.

✍ Solve the following triangular systems by hand.

**(a)**\(\displaystyle \begin{aligned} -2x_1 &= -4 \\ x_1 - x_2 &= 2 \\ 3x_1 + 2x_2 + x_3 &= 1 \end{aligned} \qquad\)**(b)**\(\displaystyle \begin{bmatrix} 4 & 0 & 0 & 0 \\ 1 & -2 & 0 & 0 \\ -1 & 4 & 4 & 0 \\ 2 & -5 & 5 & 1 \end{bmatrix} \mathbf{x} = \begin{bmatrix} -4 \\ 1 \\ -3 \\ 5 \end{bmatrix}\qquad\)**(c)**\(\displaystyle \begin{aligned} 3x_1 + 2x_2 + x_3 &= 1 \\ x_2 - x_3 &= 2 \\ 2 x_3 &= -4 \end{aligned}\)⌨ Use forwardsub to solve the systems from the previous problem. Verify that the solution is correct by computing \(\mathbf{L}\mathbf{x}\) and subtracting \(\mathbf{b}\).

⌨ Use backsub to solve the following systems. Verify that the solution is correct by computing \(\mathbf{U}\mathbf{x}\) and subtracting \(\mathbf{b}\).

**(a)**\(\displaystyle \begin{bmatrix} 3 & 1 & 0 \\ 0 & -1 & -2 \\ 0 & 0 & 3 \\ \end{bmatrix} \mathbf{x} = \begin{bmatrix} 1 \\ 1 \\ 6 \end{bmatrix}\qquad\)**(b)**\(\displaystyle \begin{bmatrix} 3 & 1 & 0 & 6 \\ 0 & -1 & -2 & 7 \\ 0 & 0 & 3 & 4 \\ 0 & 0 & 0 & 5 \end{bmatrix} \mathbf{x} = \begin{bmatrix} 4 \\ 1 \\ 1 \\ 5 \end{bmatrix}\)⌨ If \(\mathbf{B}\in\mathbb{R}^{n \times p}\), has columns \(\mathbf{b}_1,\ldots,\mathbf{b}_p\), then we can pose \(p\) linear systems at once by writing \(\mathbf{A} \mathbf{X} = \mathbf{B}\), where \(\mathbf{X}\) is \(n\times p\). Specifically, this equation implies \(\mathbf{A} \mathbf{x}_j = \mathbf{b}_j\) for \(j=1,\ldots,p\).

**(a)**Modify forwardsub and backsub so that they solve the case where the second input is \(n\times p\) for \(p\ge 1\).**(b)**If \(\mathbf{A} \mathbf{X}=\mathbf{I}\), then \(\mathbf{X}=\mathbf{A}^{-1}\). Use this fact to write a function`ltinverse`

that uses your modified**forwardsub**to compute the inverse of a lower triangular matrix. Test your function on at least two nontrivial matrices. (We remind you here that this is just an exercise; matrix inverses are rarely a good idea in numerical practice!)⌨ Solving triangular systems showed solutions of \(\mathbf{A}\mathbf{x}=\mathbf{b}\), where

\[\begin{split}\mathbf{A} = \begin{bmatrix} 1 & -1 & 0 & \alpha-\beta & \beta \\ 0 & 1 & -1 & 0 & 0 \\ 0 & 0 & 1 & -1 & 0 \\ 0 & 0 & 0 & 1 & -1 \\ 0 & 0 & 0 & 0 & 1 \end{bmatrix}, \quad \mathbf{b} = \begin{bmatrix} \alpha \\ 0 \\ 0 \\ 0 \\ 1 \end{bmatrix}.\end{split}\]Solve with \(\alpha=0.1\) and \(\beta=10,100,\ldots,10^{12}\), making a table of the values of \(\beta\) and \(|x_1-1|\).