next up previous print clean
Next: Least-squares solution of the Up: Algorithm for minimizing the Previous: An efficient algorithm for

Minimizing the Huber function

So far I have introduced a general method for solving nonlinear problems. In this section I show how this algorithm can be used when the Huber function is utilized for measuring the data misfit. In fact, we only need to derive the gradient of the objective function in equation ([*]). The gradient can be written in the following compact form Li and Swetist (1998):
\begin{displaymath}
\bold{\nabla} f(\bold{m})=\bold{L}'(\bold{Lm}-\bold{d})^\alpha_{-\alpha},\end{displaymath} (84)
where ${\bf z}^\alpha_{-\alpha}$ is a vector with ith component
\begin{displaymath}
z_i \leftarrow max\{-1,min\{1,z_i/\alpha\}\}.\end{displaymath} (85)
A last difficulty arises in the choice of the threshold $\alpha$ in equation ([*]). This value remains constant during the iterations. $\alpha$ is also the only parameter to choose a priori for different problems. I have not derived any analytical expression for $\alpha$, but based on previous works with IRLS methods Darche (1989), it seems that  
 \begin{displaymath}
\alpha = \frac{max\vert\bold{d}\vert}{100}\end{displaymath} (86)
is a good practical choice. Another possible solution is to set $\alpha$ at the 98-th percentile of the data (Claerbout, 2000, Personal communication).
next up previous print clean
Next: Least-squares solution of the Up: Algorithm for minimizing the Previous: An efficient algorithm for
Stanford Exploration Project
5/5/2005