next up [*] print clean
Next: About this document ... Up: Table of Contents

Spectral Factorization by the Wilson algorithm

John Burg (as interpreted by JFC)

jon@sep.stanford.edu

Newton's iteration for square roots  
 \begin{displaymath}
a_{t+1} \quad =\quad{1\over 2} \ \left( a_t \ +\ {s\over a_t} \right)\end{displaymath} (1)
converges quadratically starting from any real initial guess a0 except zero. When a0 is negative, Newton's iteration converges to the negative square root.

Quadratic convergence means that the square of the error $a_t-\sqrt{s}$at one iteration is proportional to the error at the next iteration
   \begin{eqnarray}
a_{t+1} - \sqrt{s} \quad\sim\quad ( a_t-\sqrt{s})^2 
\quad =\quad a_t^2 - 2 a_t \sqrt{s} + s \quad \gt \quad 0\end{eqnarray} (2)
so, for example if the error is one significant digit at one iteration, at the next iteration it is two digits, then four, etc. We cannot use equation (2) in place of the Newton iteration itself, because it uses the answer $\sqrt{s}$ to get the answer at+1, and also we need the factor of proportionality. Notice, however, if we take the factor to be 1/(2at), then $\sqrt{s}$ cancels and equation (2) becomes itself the Newton iteration (1).

Another interesting feature of the Newton iteration is that all iterations (except possibly the initial guess) are above the ultimate square root. This is obvious from equation (2).

We can insert spectral functions in the Newton square-root iteration, for example $s(\omega)$ and $a(\omega)$.Where the first guess a0 happens to match $\sqrt{s}$,it will match $\sqrt{s}$ at all iterations. Something inspires Wilson to write the Newton iteration as

\begin{displaymath}
2\ {a_{t+1}\over a_t} \quad =\quad1 \ +\ {s\over a_t^2}\end{displaymath} (3)
and then express the spectrum $S=\bar A A$ as a Z-transform.

 
 \begin{displaymath}
{\bar A_{t+1}(1/Z) \over \bar A_t(1/Z)}
\ +\ 
{A_{t+1}(Z) \over A_t(Z)}
\quad =\quad
1 \ +\ {S(Z) \over \bar A_t(1/Z)\ A_t(Z)}\end{displaymath} (4)

Now we are ready for the algorithm: Compute the right side of (4) by polynomial division forwards and backwards and then add 1. Then abandon negative lags and take half of the zero lag. Now you have At+1(Z) / At(Z). Multiply out (convolve) the denominator At(Z), and you have the desired result At+1(Z). Iterate as long as you wish.

We show that At+1(Z) is minimum phase: Both sides of (4) are positive, as noted earlier. Both terms on the right are positive. Since the Newton iteration always overestimates, the 1 dominates the rightmost term. After masking off the negative powers of Z (and half the zero power), the right side of (4) adds two wavelets. The 1/2 is wholly real, and hence its real part always dominates the real part of the rightmost term. Thus (after masking negative powers) the wavelet on the right side of (4) has a positive real part, so the phase cannot loop about the origin. This wavelet multiplies At(Z) to give the final wavelet At+1(Z) and the product of two minimum-phase wavelets is minimum phase.

Claerbout is interested in factoring the Laplacian
\begin{displaymath}
S(Z) \quad =\quad-{1\over Z^{100}} - {1\over Z} + 4 -Z -Z^{100}\end{displaymath} (5)
It seems less confusing than the Kolmogoroff method, and so easier to teach. Furthermore, it might be better because we can think of upper and lower triangular matrices, instead of recursive filters. The advantage of the triangular matrices is that they embody time-variable filters.

Burg is interested in extending the Wilson algorithm to cross spectra.

\begin{displaymath}
\bold A
\quad =\quad
\left[
\begin{array}
{cccccccccccc}
 1....
 ...t&\cdot&\cdot&\cdot& -.6& -.2& 0& -.6& 1.8\\ \end{array}\right]\end{displaymath} (6)
 



 
next up [*] print clean
Next: About this document ... Up: Table of Contents
Stanford Exploration Project
4/12/1998