The Under-specified problem
I think of the simple linear x y z problems of the previous pages as the attempt
to find the single point in 3-space that satisfies the given equations.
If the resultant square matrix is singular, then instead of finding a single point, we find a point on the line and we have
an equation we can use to find the equation of the line, if desired, using
(I - A‡A) z, or by directly examining
the unused column(s) of the C matrix
Occasionally a matrix problem arises where there are more unknowns than equations.
This happens, for instance, when combining Taylor's series for approximating the derivatives when using Finite-Differences
(as I was wont to do back before learning Finite-Elements.)
In this case, there's multiple possibilities but these are all related.
The equations
Here's a typical problem with one more unknown than equations:
The same vector notation applies, even if the A matrix is not square.
The matrix A is:
The vector x is:
The vector b is:
Inverting the non-square matrix
To solve for the x vector, we augment the matrix for row and column
operations as before, but we must use Identity matrices of the appropriate rank:
Subtract row 1 from row 2.
Divide through row 2 by -4.
Subtract 2 of column 2 from column 3.
Subtract column 1 from column 2.
Add column 2 to column 3.
Divide through row 1 by 2.
The original matrix is completely reduced.
What remains is an Identity matrix of rank 2, therefore we use
2 columns of the C matrix
and 2 rows of the R matrix.
The Pseudo-inverse of the under-specified problem
Multiply the C matrix into the R matrix to
obtain the A‡ pseudo-inverse.
Confirm the two generalized matrix properties
The first thing to notice is the A‡ pseudo-inverse is not square.
We must be able to multiply this pseudo-inverse times our original A matrix on the
right or left.
For this to work, our pseudo-inverse must have as many columns as the original had rows, and as many rows as the
original had columns.
Left and right hand inverse
Multiply our A‡ pseudo-inverse times the original
A matrix.
Then multiply the original A matrix times this one to get
A A‡ A which should equal A.
Therefore our A‡ pseudo-inverse meets the first property of
the generalized matrices discussed earlier.
It is the left-hand inverse of the original matrix.
Now check the pseudo-inverse and original for the second property - right-hand inverse.
Multiply our intermediate matrix times the pseudo-inverse.
This is A‡ A A‡
which should equal A‡.
Both properties are confirmed.
This solution might not be the "least squares" solution obtained by Moore-Penrose, but the method doesn't
suffer from the problems we saw earlier.
In the next section we look at applying this method to an even more under-specified problem.
And here's a short note on how to go directly from this work to the parametric or symmetric equation of a line.