\documentstyle[twoside,amssymb]{article}
  % amssymb is used for R in Real numbers
\pagestyle{myheadings}
\markboth{\hfil Stable evaluation of differential operators \hfil EJDE--1997/15}%
{EJDE--1997/15\hfil Otmar Scherzer\hfil}
\begin{document}
\title{\vspace{-1in}\parbox{\linewidth}{\footnotesize\noindent
{\sc  Electronic Journal of Differential Equations},
Vol.\ {\bf 1997}(1997), No.\ 15, pp. 1--12. \newline
ISSN: 1072-6691. URL: http://ejde.math.swt.edu or http://ejde.math.unt.edu
\newline ftp (login: ftp) 147.26.103.110 or 129.120.3.113}
 \vspace{\bigskipamount} \\
Stable evaluation of differential operators and linear and 
nonlinear multi--scale filtering 
\thanks{ {\em 1991 Mathematics Subject Classifications:}
65J10, 65J15,  65K10, 35J60, 26A45.\hfil\break\indent
{\em Key words and phrases:} Nondifferntiable optimization problems, 
regularization, \hfil\break\indent
inverse problems, image reconstruction, bounded variation norm.
\hfil\break\indent
\copyright 1997 Southwest Texas State University  and University of
North Texas. \hfil\break\indent
Submitted July 8, 1997. Published September 10, 1997.\hfil\break\indent
Partly supported by the Christian--Doppler--Society, Austria}}
\date{}
\author{Otmar Scherzer}
\maketitle

\begin{abstract}
Diffusion processes create multi--scale analyses, which enable
the generation of simplified pictures, where for 
increasing scale the image gets sketchier. In many practical 
applications the ``scaled image'' can be characterized via a 
variational formulation as the solution of a minimization problem 
involving unbounded operators. These unbounded operators can be 
evaluated by regularization techniques. We show that the 
theory of stable evaluation of unbounded operators can be applied
to efficiently solve these minimization problems.
\end{abstract}

\newtheorem{theorem}{Theorem}[section]
\newtheorem{example}[theorem]{Example}
\def\bea{\begin{eqnarray}}
\def\eea{\end{eqnarray}}
\def\beas{\begin{eqnarray*}}
\def\eeas{\end{eqnarray*}}
\def\be{\begin{equation}}
\def\ee{\end{equation}}

\def\aa{A_{{\lambda^2}\beta\gamma}}
\def\bv{{\mbox{BV}}}
\def\dif{\nabla .}
\def\lgs{L_{\gamma}^*}
\def\jbg{J_{\beta\gamma}}


\section{Introduction}
\label{sec1}
Morel and Solimini describe in their book \cite{MorSol95} a 
{\it{multi--scale analysis}}:
\begin{quotation}
Multi--scale analysis is concerned with the generation of a sequence of 
  simplified pictures $u_\lambda$ from a single initial picture $u_0^\delta$, 
  where $u_\lambda$ appears to be a rougher and sketchier version of 
  $u_0^\delta$ as $\lambda$ increases.
\end{quotation}
In the literature various possibilities for generating multi--scale 
analyses have been proposed. 
One common method to generate a multi--scale analysis is by a
{\it{diffusion process}}, such as
\bea
\label{1.1}
\begin{array}{rcl}
\frac{\partial u}{\partial t}(x,t) &=& \Delta u(x,t)\,\\
u(x,0) &=& u_0^\delta(x)\;.
\end{array}
\eea
It is easy to see that in ${\Bbb R}^2$ convolution of the signal $u_0^\delta$ 
at each 
scale is equivalent to the solution of the heat equation with 
the signal as initial datum \cite{Koe84}. 
The solution of (\ref{1.1}) for the initial datum $u_0^\delta$ with 
bounded quadratic norm is 
$$
u(x,t) = G_t \star u_0^\delta(x) = \int G_t(x-y)u_0^\delta(y)\,dy\;$$
where $G_\sigma(x) = \frac{1}{4\pi\sigma} \exp (-\|x\|^2/4\sigma)$ 
($x \in {\Bbb R}^2$).
The sequence $u_\lambda(.)=u(.,\lambda)$ defines a multi--scale 
analysis, which satisfies the following properties: Let $S_\lambda$
be the mapping $u_0^\delta \to S_\lambda(u_0^\delta) := u_\lambda$
\begin{enumerate}
\item $u_\lambda \to u_0^\delta$ as $\lambda \to 0$
\item $S_\lambda(u_0^\delta)$ only depends on $S_{\lambda'}(u_0^\delta)$ if 
      $\lambda > \lambda'$
\item If $A$ is an isometry, then 
      $S_\lambda(u_0^\delta \circ A) = (S_\lambda(u_0^\delta)) \circ A$.
\end{enumerate}
In image processing and image enhancement techniques one easily 
understands the necessity of a previous low--pass filtering, 
i.e., by convolution of the signal with a Gaussian kernel:
If the signal $u_0^\delta$ is noisy, the gradient has a lot of irrelevant 
spikes, which have to be eliminated.

As we will see in Section \ref{sec3} the linear multi--scale analysis 
based on (\ref{1.1}) has a tendency to over smooth data.
Many attempts have been made in the literature to construct diffusion
processes which satisfy a {\it{strong causality}} principle, that 
is, the associated multi--scale analysis reconstructs
edges and corners of images, while simultaneously eliminating
highly oscillating noise in the image. Diffusion processes where 
the associated multi--scale analysis satisfies a strong causality 
principle have been found to be mostly nonlinear. 
In Section \ref{sec2} we give an outline of linear and nonlinear 
diffusion processes and motivate associated multi--scale analysis. 
The multi--scale analysis can be the basis for efficient algorithms 
for denoising images and edge--detection. In Section \ref{sec3} 
we will outline the concept of stable evaluation of unbounded 
operators and we show in Sections \ref{sec3}, \ref{sec4} that 
this concept is relevant for the multi--scale analysis associated 
with diffusion processes. 
In Section \ref{sec5} we will discuss the numerical 
implementation of a multi--scale analysis associated with a nonlinear 
diffusion process and we will show some numerical results.

\section{Variational formulations of diffusion processes} 
\label{sec2}
\subsection{The Hildreth--Marr theory}
The {\it energy functional} associated with the differential equation 
(\ref{1.1}) is 
$$ E(u) = \int | \nabla u |^2\,.$$
This functional gives an evaluation of the amount of information in $u$. 
The energy semi norm is the right functional to be associated with 
(\ref{1.1}) since the heat equation is the descent method associated 
with this energy. If we discretize (\ref{1.1}) we end up with 
$$u(x,t) - u_0^\delta(x) = t \Delta u(x,t)\;$$
which is the solution of a variational problem: roughly speaking, 
numerically solving the heat equation is therefore equivalent to 
minimizing the following functional iteratively with 
$t=\lambda^2$ small:
$$E_\lambda(u) = \lambda^2 \int_\Omega |\nabla u|^2 + \int_\Omega (u-u_0^\delta)^2.$$
In this formulation the scaling parameter appears explicitly as the 
multiplier of the smoothing term. The solution $u_\lambda$ of this 
problem can be defined as the ``analysis of $u$ at scale $\lambda$'' 
as well as the solution of the heat equation at time $t=\lambda^2$.

\subsection{The Perona and Malik theory}
For an edge--detection filtering technique Perona and Malik 
(see e.g. \cite{MorSol95}) suggested 
the {\it nonlinear} diffusion process
\bea
\label{2.1}
\begin{array}{rcl}
\frac{\partial u}{\partial t}(x,t) &=& 
      \nabla. \left( f(|\nabla u|) \nabla u \right)(x,t)\,\\
u(x,0) &=& u_0^\delta(x)\;,
\end{array}
\eea
where $f$ is a smooth non-increasing function satisfying
$$f(0)=1, \quad f(s) \geq 0 \quad \mbox{ and } 
  \lim_{s \to \infty} f(s) \to 0\,.$$
The idea of using this type of nonlinear filtering technique is that 
the smoothing process is conditional: if $\nabla u(x)$ is large, then 
the diffusion will be low and therefore the exact location of the edges 
in the image are preserved. If $\nabla u(x)$ is small, then the diffusion will tend 
to smooth around $x$. Similarly to the Hildreth--Marr theory 
one can derive a variational formulation of the Perona and Malik model.
In fact, if we set $\tilde{f}(a) = f (\sqrt{a})$ then equation  (\ref{2.1}) can be 
rewritten as 
\beas
\begin{array}{rcl}
\frac{\partial u}{\partial t}(x,t) &=& 
      \nabla. \left( \tilde{f}(|\nabla u|^2) \nabla u \right)(x,t)\,\\
u(x,0) &=& u_0^\delta(x)\;,
\end{array}
\eeas
which yields a multi--scale analysis
\be
\label{2.2}
E_\lambda(u) = \lambda^2 \int_\Omega \tilde{F}(|\nabla u|^2)\, + 
               \int_\Omega (u-u_0^\delta)^2\,,         
\ee
where $\tilde{F}(t) = \int_0^t \tilde{f}(s)\,.$

\subsection{A nonlinear filtering technique studied by Osher and Rudin}
Based on the Perona and Malik theory Osher and Rudin \cite{OshRud90} 
considered a nonlinear diffusion process for edge detection:
\bea
\label{2.3}
\begin{array}{rcl}
\frac{\partial u}{\partial t}(x,t) &=& 
      \nabla. \left( \frac{ \nabla u } {|\nabla u|} \right)(x,t)\\
u(x,0) &=& u_0^\delta(x)\;.
\end{array}
\eea
The energy functional associated with this diffusion process is
$$E(u) = \int_\Omega |\nabla u|\;.$$
As above an associated multi--scale analysis can be motivated in terms of the functional:
\be
\label{2.4}
E_\lambda(u) = \lambda^2 \int_\Omega |\nabla u|\; + \;
                 \int_\Omega (u-u_0^\delta)^2\;.
\ee

\subsection{Higher order nonlinear filtering techniques}
Diffusion filters of higher order play an important role in parameter 
estimation problems \cite{Sch98}. An appropriate filtering 
technique for enhancing jumps in derivatives of signals is
\be
\label{2.5}
\frac{\partial u}{\partial t}(x,t) = H. \left( \frac{Hu}{|Hu|} \right)(x,t)\;.
\ee
Here $Hu$ denotes the Hessian of $u$ and $H. = (\dif \dif)$.
The multi--scale analysis associated with this diffusion process is 
$$E_\lambda(u) = \lambda^2 \int_\Omega |H u|\; + \;
                 \int_\Omega (u-u_0^\delta)^2\;.$$

\section{Unbounded operators}
\label{sec3}
In this section we consider the problem of computing values of an 
unbounded operator $L$. Here and in the following of this section, 
we will always denote by $L:{\cal{D}}(L) \subseteq H_1 \to H_2$ a 
closed densely defined unbounded linear operator between two Hilbert 
spaces $H_1$ and $H_2$. The problem of computing values $y = Lu_0$, 
for $u_0 \in D(L)$, is then ill--posed in the sense that small 
perturbations in $u_0$ may lead to data $u_0^\delta$ satisfying 
$\|u_0-u_0^\delta\| \leq \delta$, but $u_0^\delta \notin {\cal{D}}(L)$, or 
even if $u_0^\delta \in {\cal{D}}(L)$, it may happen that $Lu_0^\delta 
\not\to Lu_0$ as $\delta \to 0$. 
Morozov has studied a stable method for approximating the value $Lu_0$
when only approximate data $u_0^\delta$ is available (see \cite{Mor84}
for information on Morozov's work). This method takes as an 
approximation to $y = Lu_0$ the vector 
$$y_{\lambda^2}^\delta = Lz_{\lambda^2}^\delta,$$
where $z_{\lambda^2}^\delta$ minimizes the functional
$$\|z - u_0^\delta\|^2 + {\lambda^2} \|Lz\|^2 \quad ({\lambda^2} > 0)$$
over $D(L)$. For $L = \nabla$ the approximate $z_{\lambda^2}^\delta$ of $u_0$ 
is the solution of the Hildreth--Marr problem at level ${\lambda^2}$. 
In \cite{GroSch93} we discussed the {\it{optimal}} choice of the 
scaling parameter ${\lambda^2}$ in relation to the noise level $\delta$ 
and smoothness assumptions on the solution and the numerical realization.
Of particular interest for image reconstruction and image enhancement
techniques is the choice of the scaling parameter. The theory developed 
for the stable evaluation of unbounded operators to optimally choose 
the {\it{regularization}} parameter is {\bf directly} applicable to 
optimally choosing the {\it{scaling}} parameter in the Hildreth--Marr theory.
In order to clarify the choice of the scaling parameter we cite a result
developed in \cite{GroSch93}.
\begin{theorem}
If $u_0 \in {\cal{D}}(L^*L)$ and $u_0 \not\in {\cal{N}}(L)$, then 
$\|y_{{\lambda^2}(\delta)}^\delta - Lu_0\| = O(\sqrt{\delta})\,,$ 
when ${\lambda^2}(\delta)$ is chosen such that 
\be
\label{3.1}
\|u_{{\lambda^2}(\delta)}^\delta - u_0\| = O(\sqrt{\delta})\;.
\ee
\end{theorem}

Roughly speaking in the particular case $L = \nabla$ the assumption 
$u_0 \in {\cal{D}}(L^*L)$ means $u_0 \in H^2$, and the derivative of $u_0$ 
can be reconstructed with a rate $O(\sqrt{\delta})$ if ${\lambda^2}(\delta)$ 
is chosen according to the discrepancy principle (\ref{3.1}).

Other strategies for optimally choosing the regularization parameters 
when calculating unbounded operators can be found in \cite{GroSch93}.
For more background on the stable evaluation of unbounded operators we 
refer to \cite{Gro92}.

From a simple example, by taking $L=\nabla$, ${\cal{D}}(L) = H^1({\Bbb R})$, 
the Sobolev space of functions possessing a weak derivative in $L^2({\Bbb R})$, 
it follows that $z_{\lambda^2}^\delta \in H^1({\Bbb R})$. 
This shows that at any scale discontinuities of the original image are 
smeared out, and thus the method is not strongly causal. Nonlinear filtering 
techniques have turned out to be strongly causal and therefore, we study 
below numerical implementation of nonlinear filtering techniques based on 
the nonlinear diffusion process developed by Rudin, Osher, and Fatemi 
\cite{RudOshFat92}.

\section{Unbounded operators in nonlinear multi--scale analysis}
\label{sec4}
The multi--scale analysis associated with the nonlinear diffusion process 
developed by Osher and Rudin is sometimes called 
{\it{regularization with functions of bounded variation}}, 
since the scale--energy functional (\ref{2.4}) is finite for any 
function $u \in \bv$, that is, the space of all functions which satisfy
$$\int\limits_\Omega |\nabla u| := 
        \sup \left\{ \int\limits_\Omega u \nabla. v dx\,:\,
                     v \in C_0^\infty(\Omega,{\Bbb R}^d), |v(x)| \leq 1, 
                     x \in \Omega \right\} < \infty\;.$$
In practical applications, since the functional (\ref{2.4}) is not 
differentiable, the minimizer of (\ref{2.4}) is frequently approximated 
by the solution of the 
(nonlinear) {\it{least squares problem}} (see e.g. 
\cite{DobVog95,DobSan94,DobSch96}) 
\be
\label{4.1}
\min_{u \in L^2(\Omega)} \left\{ \frac{1}{2} \|u-u_0^\delta\|^2_{L^2(\Omega)} + 
                          {\lambda^2} J_\beta(u)
                  \right\}\,,
\ee
where 
\be
\label{4.2}
J_\beta(u):=\int_{\Omega} \sqrt{ |\nabla u|^2 + \beta^2}\;;
\ee
here $\Omega \subseteq {\Bbb R}^d$. The term $J_\beta(u)$ is used 
for stabilization and is therefore called the {\it{regularizing term}}. 

Vogel and Oman \cite{VogOma96} introduced a fixed point iteration method
to minimize (\ref{4.1}), (\ref{4.2}): 
The corresponding (formally derived) {\it{Euler--Lagrange}} equation is
$$
u - {\lambda^2} \nabla . 
    \left( \frac{1}{\sqrt{|\nabla u|^2 + \beta^2}} \nabla u\right) 
    = u_0^\delta\;,
$$
which can be expressed in operator notation as
\be
\label{4.3}
(I + A_{{\lambda^2}\beta}(u)) u = u_0^\delta,
\ee
where 
$$
A_{{\lambda^2}\beta}(u)v := - {\lambda^2} \nabla . 
    \left( \frac{1}{\sqrt{|\nabla u|^2 + \beta^2}} \nabla v\right)\;.$$
A fixed point iteration for (\ref{4.3}) can be defined by 
$$
u^{n+1} = (I + A_{{\lambda^2}\beta}(u^n))^{-1} u_0^\delta =: F(u^n), 
\quad n=0,1,...$$
In actual computations the minimization of (\ref{4.1}) is difficult 
to handle, since it involves evaluations of the {\it{unbounded}} 
operator $\nabla$. To overcome this difficulty we proposed in 
\cite{DobSch96} to approximate the values of the unbounded operator 
$\nabla$ and use the minimizing element of  
\be
\label{4.4}
\min_{u \in L^2(\Omega)} \left\{ \frac{1}{2} \|u-u_0^\delta\|^2_{L^2(\Omega)} 
+ {\lambda^2} \jbg(u)
\right\}\;,
\ee
where 
\be
\label{4.5}
\jbg(u):=\int_{\Omega} \sqrt{ |L_\gamma u|^2 + \beta^2},
\ee
as an approximation for the desired image. In practical applications 
the operator $L_\gamma$ might be, for instance, a finite difference 
scheme, where the parameter $\gamma$ represents the grid-size.
Denoting by $L_\gamma^*$ the adjoint of $L_\gamma$, the Euler--Lagrange 
equation for a minimizer of (\ref{4.4}) is 
$$
u + {\lambda^2} \lgs 
\left( \frac{1}{\sqrt{|L_\gamma u |^2 + \beta^2}} L_\gamma u \right) = 
u_0^\delta\;.
$$
Extending the ideas of Vogel and Oman \cite{VogOma96} and Dobson and Vogel
\cite{DobVog95} we studied (in \cite{DobSch96}) the iterative scheme
\be
\label{4.6}
u^{n+1} = (I+\aa(u^n))^{-1}u_0^\delta =: F(u^n),
\ee
where 
\be
\label{4.7}
\aa(u)v:= \frac{{\lambda^2}}{\beta} \lgs
\left( \frac{1}{\sqrt{\frac{|L_\gamma u |^2}{\beta^2} + 1}} L_\gamma v\right).
\ee
As long as the parameters ${\lambda^2}, \beta, \gamma$ are chosen 
appropriately (\ref{4.6}),(\ref{4.7}) is convergent. This 
setting reflects computational relevance since in numerical 
simulations the operator $\nabla$ is 
(almost) never calculated exactly, but approximately evaluated, 
e.g. by some finite difference scheme.
 
In numerical calculations (see e.g. \cite{DobSan94,ChaLio95}) 
regularization with functions of bounded variation turned out 
to be very efficient to determine discontinuities in signals 
and images. This can be motivated both by arguing that the 
associated diffusion process (\ref{2.3}) is designed to optimally 
enhance edges in images (see e.g. \cite{OshRud90,MorSol95})
or from arguing for the least squares functional (\ref{4.1}) 
(respectively (\ref{2.4})) directly. Namely,
efficient regularizing schemes have to satisfy the following 
two properties:
\begin{enumerate}
\item The regularizing term evaluated at the function to be 
      recovered is finite; 
      in signal and image processing, the images of most interest 
      are piecewise constant and thus have finite bounded variation 
      (see e.g. \cite{EvaGar92,Zie80}), i.e., 
      $$\int\limits_\Omega |\nabla u_0| := 
        \sup \left\{ \int\limits_\Omega u_0 \nabla. v dx\,:\,
                     v \in C_0^\infty(\Omega,{\Bbb R}^d), |v(x)| \leq 1, 
                     x \in \Omega \right\} < \infty\;.$$
\item The recovery of images from noisy measurements is ill--posed, 
      in the sense that small perturbations in the data $u_0^\delta$ may 
      lead to significant distortions of the picture
      (mathematically the ill-posedness is due to the fact that the 
      bounded variation semi norm -- which represents a natural norm 
      in measuring the distortion of the picture -- does not depend 
      continuously on the $L^2$ -- norm -- which represents the natural 
      norm for noise in the data), the regularizing term has to 
      have appropriate stabilizing properties. 
\end{enumerate}

In practical applications (such as in non--destructive testing) 
frequently discontinuities in derivatives have to be recovered 
instead of the function itself. Examples include electrical 
conductivity problems in ${\Bbb R}^1,{\Bbb R}^2$, where the idealized model 
is the following: Given some domain $\Omega$ in ${\Bbb R}^1$ or ${\Bbb R}^2$; 
the voltage potential $u_0$ inside of $\Omega$ satisfies 
\be
\label{4.8}
\nabla .(\sigma \nabla u_0) = 0 \quad \mbox{ in } \Omega.
\ee 
If the electrical conductivity $\sigma$ is discontinuous, then by 
jump relations (see e.g. \cite{Isa90}) the resulting solution of 
(\ref{4.8}), the voltage potential $u_0$, reveals discontinuities 
in the derivatives of $u_0$. Therefore, discontinuities in the 
material (i.e., discontinuities in $\sigma$) 
can be detected from measurement data $u_0^\delta$ of $u_0$ 
and by {\it{denoising}} $u_0^\delta$ and detecting discontinuities 
in derivatives of $u_0$.  

For estimating discontinuities in electrical conductivity problems by 
denoising the data $u_0^\delta$ the two demands on an efficient regularization 
scheme suggest the use a regularizing (semi) norm which satisfies the 
two properties:
\begin{enumerate}
\item For functions with jumps in derivatives the regularizing norm is finite
\item The stabilizing term has to cope with the ill--posedness.
\end{enumerate}
A reasonable semi norm satisfying both properties is 
$$
|Hu| = \sum_{i=1}^d \sum_{j=1}^d 
       \left| \frac{\partial^2 u}{\partial x_i \partial x_j} \right|;
$$
here $Hu$ denotes the {\it{Hessian}} of $u$, i.e.,
\beas
H = \left( 
\begin{array}{rcl}
\frac{\partial^2 u}{\partial x_1^2}  & .... & 
\frac{\partial^2 u}{\partial x_1 \partial x_d} \\
... & ... & ...\\
\frac{\partial^2 u}{\partial x_d \partial x_1} & .... & 
\frac{\partial^2 u}{\partial x_d^2}
\end{array}
\right).
\eeas
For the numerical solution of the parameter estimation 
problem described in (\ref{4.8}) we propose to use a  
{\it{higher order bounded variation denoising}} algorithm
\be
\label{4.9}
\min_{u \in L^2(\Omega)} \frac{1}{2} 
\|u-u_0^\delta\|^2_{L^2(\Omega)} + {\lambda^2} \jbg(u),
\ee
where 
$$\jbg(u):=\int_{\Omega} \sqrt{ |H_\gamma u|^2 + \beta^2}\,;$$
here the meaning of $H_\gamma$ is a stable approximation of the Hessian 
in ${\Bbb R}^d$.

Formally (\ref{4.9}) looks like (\ref{4.5}), and actually 
the analysis of the fixed point iteration is formally the same (see 
\cite{Sch98}).

\section{Numerical results: Higher order bounded variation regularization 
           in comparison to a linear multi--scale algorithm}
\label{sec5}
In this section we present some numerical experiments to 
estimate discontinuities in derivatives of a one-dimensional 
signal from its noisy measurements.
We solved the denoising problem both with regularization
with {\it{higher order bounded variations}} and with a 
{\it linear multi--scale algorithm} and compared the numerical 
results.

In the numerical experiments presented below, when functions 
with derivatives of bounded variation were used for regularization, 
we applied the iterative scheme (\ref{4.9}) in a discrete setting. 
\begin{example}
\label{ex5.1}
{\rm 
{\bf{Regularization with functions with derivatives of higher order 
          bounded variations:}}
The first example is to determine the discontinuity 
in the derivative of the piecewise continuously differentiable function  
\bea
\label{5.1}
\begin{array}{rcl}
f : [0,1]  &\to& {\Bbb R}\,.\\
     x     &\to& \frac{1}{2}-\left| \frac{1}{2} - x \right|
\end{array}
\eea
In Figure \ref{fig5.1} the unperturbed signal and its derivative 
are plotted. 
We added to $97$ uniformly distributed measurement points random noise 
to the data (see Figure \ref{fig5.2}) to simulate noisy measurements of 
the signal. 
\begin{figure}[ht]
\unitlength1cm
\begin{minipage}[b]{6.0cm}
\begin{picture}(4.0,4.5)
\put(-0.75,-3.0){\special{psfile=st_01.eps hscale=35 vscale=35}}
\end{picture}
\par
{\caption{\label{fig5.1}}}
\end{minipage}
\hfill
\begin{minipage}[b]{6.0cm}
\begin{picture}(4.0,4.5)
\put(-0.75,-3.0){\special{psfile=st_02.eps hscale=35 vscale=35}}
\end{picture}
\par
{\caption{\label{fig5.2}}}
\end{minipage}
\end{figure}
Figures \ref{fig5.3}, \ref{fig5.4} show reconstructions with  
regularization with {\it{higher order derivatives of bounded 
variation}}. 
\begin{figure}[ht]
\unitlength1cm
\begin{minipage}[b]{6.0cm}
\begin{picture}(4.0,4.5)
\put(-0.75,-3.0){\special{psfile=st_03.eps hscale=35 vscale=35}}
\end{picture}
\par
{\tiny{\caption{\label{fig5.3} ${\lambda^2} = 1.e-4$, 
                              $\beta = 1.e-2$,
                              $\gamma = 1/m$, $m=96$}}}
\end{minipage}
\hfill
\begin{minipage}[b]{6.0cm}
\begin{picture}(4.0,4.5)
\put(-0.75,-3.0){\special{psfile=st_04.eps hscale=35 vscale=35}}
\end{picture}
\par
{\tiny{\caption{\label{fig5.4} ${\lambda^2} = 1.e-3$, 
                         $\beta = 1.e-2$, $\gamma = 1/m$, $m=96$}}}
\end{minipage}
\end{figure}
In each of the Figures \ref{fig5.3}, \ref{fig5.4} 
the dotted line represents the reconstructed function, while the dashed 
dotted line represents its derivative, which was calculated by a forward 
difference scheme.
In the bottom line of these figures the actual parameter setting of 
$\lambda^2, \beta, \gamma$ is plotted. 
}
\end{example}

\begin{example}
\label{ex5.2}
{\rm {\bf{A linear multi--scale algorithm:}}
Here the same denoising problem as in Example \ref{ex5.1} is studied, 
but as resolving algorithm we used a linear multi--scale algorithm.
In Figures \ref{fig5.5}, \ref{fig5.6} we have plotted the Tikhonov 
regularized solutions, i.e., the minimizing elements of 
\be
\label{5.2}
\min \|u - u_0^\delta\|^2_{L^2} + {\lambda^2} \|u_{xx}\|^2_{L^2}
\ee
for different values of the regularization parameter ${\lambda^2}$. 
The reconstruction in Figures \ref{fig5.5} - \ref{fig5.6} the 
$\star$--line shows reconstructions with a linear multi--scale algorithm if it is 
implemented with noisy boundary data, and the ``$\times$''--line shows its 
derivative; lines $-\cdot$ and $--$ denote the reconstruction and its 
derivative if ``a--priori'' information on exact boundary data is 
available and implemented.
\begin{figure}[ht]
\unitlength1cm
\begin{minipage}[b]{6.0cm}
\begin{picture}(4.0,4.5)
\put(-0.75,-3.0){\special{psfile=st_05.eps hscale=35 vscale=35}}
\end{picture}
\par
{\tiny{\caption{\label{fig5.5} ${\lambda^2} = 1.e-4$, $m=96$}}}
\end{minipage}
\hfill
\begin{minipage}[b]{6.0cm}
\begin{picture}(4.0,4.5)
\put(-0.75,-3.0){\special{psfile=st_06.eps hscale=35 vscale=35}}
\end{picture}
\par
{\tiny{\caption{\label{fig5.6} ${\lambda^2} = 1.e-3$, $m=96$}}}
\end{minipage}
\end{figure}
}
\end{example} 
It is apparent from the numerical experiments that regularization 
with higher order bounded variation resolves discontinuities in derivatives 
efficiently; in comparison with a linear multi--scale algorithm the location 
of the discontinuities can be sharply estimated.
Of course there is a price to pay: a linear multi--scale algorithm requires 
only the solution of one sparse matrix problem. For regularization with functions of higher order bounded variation 
a sparse matrix problem (of the same dimension as in a linear multi--scale algorithm) 
has to be solved in each iteration. 

\begin{thebibliography}{10}

\bibitem{ChaLio95}
A.~Chambolle and P.-L. Lions.
\newblock Image recovery via total variation minimization and related problems.
\newblock {\em Numer. Math.}, 72:168-188, 1997.

\bibitem{DobSan94}
D.~Dobson and F.~Santosa.
\newblock An image enhancement technique for electrical impedance tomography.
\newblock {\em Inverse Problems}, 10:317--334, 1994.

\bibitem{DobSch96}
D.~Dobson and O.~Scherzer.
\newblock Analysis of regularized total variation penalty methods for
  denoising.
\newblock {\em Inverse Problems}, 12:601--617, 1996.

\bibitem{DobVog95}
D.~Dobson and C.R. Vogel.
\newblock Convergence of an iterative method for total variation denoising.
\newblock {\em SIAM J. on Numerical Analysis}.
\newblock to appear.

\bibitem{EvaGar92}
L.~C. Evans and R.~F. Gariepy.
\newblock {\em Measure Theory and Fine Properties of Functions}.
\newblock CRC Press, Ann Arbor, 1995.

\bibitem{Gro92}
C.~W. Groetsch.
\newblock Spectral methods for linear inverse problems with unbounded
  operators.
\newblock {\em J. Approx. Th.}, 70:16--28, 1992.

\bibitem{GroSch93}
C.~W. Groetsch and O.~Scherzer.
\newblock Optimal order of convergence for stable evaluation of differential
  operators.
\newblock {\em Electronic Journal of Differential Equations}, 1993 No.4:1--10,
  1993.

\bibitem{Isa90}
V.~Isakov.
\newblock {\em Inverse Source Problems}.
\newblock Amer. Math. Soc., Providence, 1990.

\bibitem{Koe84}
J.~J. Koenderink.
\newblock The structure of images.
\newblock {\em Biol. Cybern.}, 50:363--370, 1984.

\bibitem{MorSol95}
J.-M. Morel and S.~Solimini.
\newblock {\em Variational Methods in Image Segmentation}.
\newblock Birkh\"auser, Basel, 1995.

\bibitem{Mor84}
V.~A. Morozov.
\newblock {\em Methods for solving incorrectly posed problems}.
\newblock Springer Verlag, New York, Berlin, Heidelberg, 1984.

\bibitem{OshRud90}
S.~Osher and L.~Rudin.
\newblock Feature oriented image enhancement using shock filters.
\newblock {\em SIAM J. on Numerical Analysis}, 27:919--940, 1990.

\bibitem{RudOshFat92}
L.I. Rudin, S.~Osher, and E.~Fatemi.
\newblock Nonlinear total variation based noise removal algorithms.
\newblock {\em Physica D}, 60:259--268, 1992.

\bibitem{Sch98}
O.~Scherzer.
\newblock Denoising with higher order derivatives of bounded variation and an
  application to parameter estimation.
\newblock Accepted for publication in Computing.

\bibitem{VogOma96}
C.~Vogel and M.~Oman.
\newblock Iterative methods for total variation denoising.
\newblock {\em SIAM J. Stat. Sci. Comput}, 17:227--238, 1996.

\bibitem{Zie80}
W.~P. Ziemer.
\newblock {\em Weakly Differentiable Functions}.
\newblock Springer Verlag, New York, Berlin, Heidelberg, 1980.

\end{thebibliography}


{\sc Otmar Scherzer\newline
Institut f\"ur Industriemathematik\newline
Universit\"at Linz\newline
Altenberger Str. 69,
A--4040 Linz, Austria}\newline
E-Mail address: scherzer@indmath.uni-linz.ac.at

\end{document}

