The null space is any solution that produces no data. You can add an arbitrary amount of the null space getting another solution as good as the first. Here is the full solution.
Iterative methods can neither subtract nor add any null space to your initial solution. It is obvious in this simple case, because the gradient (here the matrix adjoint) dotted into the null-space vector vanishes. Suppose and are matrices, while , , and are vectors. Although more complicated, something similar happens. You can test if an application involves a null space by comparing the results of various starting solutions.
Other traps arise in the world of images. Rarely are we able to iterate to full completion, so we might say, ``practically speaking, this application has null spaces.'' For example, if we know that zero frequency is theoretically a null space, we would say, ``The null space contains low frequencies.'' We cannot avoid such issues.
The textbook way of dealing with null spaces is to require the researcher to set up model styling goals (regularizations). Finding such goals demands assumptions from the researcher, assumptions that are often hard to specify. Luckily, there is another path to consider. Thinking more like a physicist, we could choose the initial solution more carefully.
In regression (47) extended to images, we might hope not to have a null-space problem when we begin iterating from , but this is not true. It is a pitfall, which in an application context, took me some years to recognize. Notice what happens the first step you move away from . Your solution becomes a constant times the gradient. The image extension of (47) being:
I have no general method for you,
but offer a suggestion that works for one family of applications
and may be suggestive for others.
Traditionally, it might happen
which happens when the data is better explained by
alone than by
Solve first for
Now define a new variable
Introducing your innovative concept (estimating
) your regression becomes: