\documentclass[reqno]{amsart}
\usepackage{hyperref}
\usepackage{graphicx}

\AtBeginDocument{{\noindent\small
Variational and Topological Methods:
Theory, Applications, Numerical Simulations, and Open Problems (2012).
{\em Electronic Journal of Differential Equations},
Conference 21 (2014),  pp. 173--181.
ISSN: 1072-6691.  http://ejde.math.txstate.edu,
http://ejde.math.unt.edu \newline ftp ejde.math.txstate.edu}
\thanks{\copyright 2014 Texas State University - San Marcos.}
\vspace{9mm}}

\begin{document} \setcounter{page}{173}
\title[\hfilneg EJDE-2014/Conf/21 \hfil Tikhonov regularization]
{Tikhonov regularization using Sobolev metrics}

\author[P. Kazemi, R. J. Renka \hfil EJDE-2014/Conf/21\hfilneg]
{Parimah Kazemi, Robert J. Renka}  % in alphabetical order

\address{Parimah Kazemi\newline
Department of Mathematics and Computer Science,
Ripon College, P. O. Box 248, Ripon, WI 54971-0248, USA}
\email{parimah.kazemi@gmail.com}

\address{Robert J. Renka \newline
Department of Computer Science \& Engineering,
University of North Texas, Denton, TX 76203-1366, USA}
\email{robert.renka@unt.edu}

\thanks{Published February 10, 2014.}
\subjclass[2000]{47A52, 65D25, 65F22}
\keywords{Gradient system; Ill-posed problem; least squares; 
\hfill\break\indent Sobolev gradient;
Tikhonov regularization}

\begin{abstract}
 Given an ill-posed linear operator equation $Au = f$ in a Hilbert
 space, we formulate a variational problem using Tikhonov regularization
 with a Sobolev norm of $u$, and we treat the variational problem by a
 Sobolev gradient flow.
 We show that the gradient system has a unique global solution for which
 the asymptotic limit exists with convergence in the strong sense using
 the Sobolev norm, and that the variational problem therefore has a unique
 global solution.
 We present results of numerical experiments that demonstrates the benefits
 of using a Sobolev norm for the regularizing term.
\end{abstract}

\maketitle
\numberwithin{equation}{section}
\newtheorem{theorem}{Theorem}[section]
\newtheorem{lemma}[theorem]{Lemma}
\newtheorem{proposition}[theorem]{Proposition}
%\newtheorem{corollary}[theorem]{Corollary}
\newtheorem{definition}[theorem]{Definition}
%\newtheorem{example}[theorem]{Example}
%\newtheorem{remark}[theorem]{Remark}
\allowdisplaybreaks


\section{Introduction}
Consider the following operator equation in which $A$ is a linear mapping from
a Hilbert space $L$ to another Hilbert space $K$:
\begin{equation}\label{e1}
   Au = f.
\end{equation}
The equation is said to be well-posed if it satisfies the properties of
existence, uniqueness, and stability; i.e., a solution $u$ exists, it is
unique, and it depends continuously on the data $f$ ($A$ has bounded
inverse).  Failure to satisfy any of the three properties characterizes the
problem as ill-posed.  These terms originated in the context of differential
equations \cite{hadamard02}, but have been applied to problems, both linear
and nonlinear, in almost every area of mathematics \cite{kabanikhin08}.
Ill-posed operator equations with $A$ unbounded are relatively rare but
have received some attention recently in \cite{ramm08} and \cite{hofmann09}.
Our focus will be on problems in which $A$ is injective and compact.

To treat \eqref{e1} when it is ill-posed, we can seek a
minimizer of the least squares functional
\begin{equation*}
   \phi(u) = \frac{1}{2} \|Au - f\|_K^2.
\end{equation*}
However, if \eqref{e1} has a nonunique solution so that $A$ has a nontrivial
kernel, then the minimization problem will not have a unique solution.  We
may choose the minimum-norm solution in this case.  If
$A$ is invertible but $A^{-1}$ is not continuous, then any noise present
in $f$ can lead to an arbitrarily large change in $u$.  In this case
we must regularize the problem in order to obtain a stable solution.  One
method is Tikhonov regularization in which we balance the size of the
residual $\phi(u)$ against the norm of the solution $u$ by minimizing
\begin{equation}\label{e2}
   \phi_{\alpha} (u) = \frac{1}{2}\|Au - f\|^2 +
                       \frac{\alpha}{2} \|u\|^2
\end{equation}
for positive regularization parameter $\alpha$
\cite{tikhonov63, tikhonov77, tikhonov98}.  We thus have a convex
minimization problem.  Using the $L^2$ norm in \eqref{e2}, the solution is
$u = (\alpha I + A^*A)^{-1}A^*f$
when $A$ is bounded and $f$ is in the domain of $A^*$.

Tikhonov regularization with $A$ compact and injective has been extensively
studied and is well understood, but the regularization term is almost always
formulated in terms of the $L^2$ norm or the Euclidean norm in the finite
dimensional case.  The primary focus of this work is to demonstrate the
advantage of using a discretized Sobolev norm for $u$.
This has been referred to as higher-order Tikhonov regularization in
\cite{aster13}.  Our contribution is an experiment that demonstrates
the effectiveness of the Sobolev norm approach for numerical differentiation,
and an analysis that includes convergence of the gradient flow and
expressions for the gradient and solution of the minimization problem.
The regularity properties of the solution evince the
appropriateness of regularization with a Sobolev norm in place of the
$L^2$ norm.

In Sections 2, 3, and 4 we develop the analysis
for the problem of minimizing $\phi_{\alpha}$ using a Sobolev gradient
flow.  We show that the gradient system has a global solution, the
solution is unique, it has an asymptotic limit, and
convergence to the limit is in the strong sense using the Sobolev norm.
We derive an expression for the regularizing operator and for the solution
$u_{\alpha}$ of the minimization problem.
In Section 5 we present results of a numerical test
that demonstrates the superiority of the Sobolev-norm based
regularization strategy when $u$ is expected to be smooth.


\section{Tikhonov regularization in Hilbert space}

Suppose that $L$ and $K$ are $L^2$ spaces, $H$ is a Sobolev space compactly
and densely embedded in $L$, and $A: L \to K$ is a bounded linear
operator.  We wish to minimize the functional
$\phi_{\alpha}:H \to \mathbb{R}$ defined by
\begin{equation*}
   \phi_{\alpha} (u) = \frac{1}{2}\|Au - f\|_K^2 +
                       \frac{\alpha}{2} \|u\|_H^2,
\end{equation*}
where $\alpha > 0$ and we have used the stronger $H$ norm in place of the
more commonly used $L$ norm.  The first and second Fr\'{e}chet derivatives are
\begin{equation}\label{e3}
  \phi_{\alpha}'(u)h = \langle Ah,Au-f \rangle_K + \alpha \langle h,u \rangle_H
\end{equation}
and
\begin{equation*}
  \phi_{\alpha}''(u)(h,h) = \| Ah\|_K^2 + \alpha \|h\|_H^2
\end{equation*}
for $h \in H$.  Note that $\phi_{\alpha}$ is an everywhere defined $C^2$
function defined on the subspace $H$, and satisfies the following definition
of convexity.

\begin{definition}\label{convexdef} \rm
Suppose $F$ is a $C^2$ function defined on a Hilbert space $H$.  $F$ is said
to be convex if there exists a positive number $\epsilon$ so that for all
$h \in H$,
\begin{equation*}
  F''(u)(h,h) \geq \epsilon \|h\|^2_H.
\end{equation*}
\end{definition}


\section{Minimization using Sobolev gradients}
To obtain a minimum of $\phi_{\alpha}$, we compute a gradient with
respect to the metric
%
\begin{equation*}
  \langle v , w \rangle_{\alpha, H, A} =  \langle v , w \rangle_H +
  \frac{1}{\alpha}\langle Av , Aw \rangle_K.
\end{equation*}
%
The following definition of a Sobolev gradient is taken from
\cite{neuberger2010}.

\begin{definition} \rm
Suppose that $F$ is a real or complex valued Fr\'{e}chet differentiable
function that is everywhere defined on a Hilbert space $H$.  Then for each
$u \in H$, there exists a unique element $\nabla_H F(u) \in H$ such that
\begin{equation*}
  F'(u)h = \langle h , \nabla_H F(u) \rangle_H
\end{equation*}
for all $h \in H$.
The gradient system associated with $F$ is
%
\begin{equation}\label{flowgen}
  z(0)=u_0 \text{ and } z'(t) = - \nabla_H F (z(t))
\end{equation}
%
for $u_0 \in H$.
\end{definition}

Now suppose that $H$ is a Sobolev space of order $k \geq 1$.
Then
\[
\langle v,w \rangle_H = \left\langle \binom {v}{Dv} , \binom{w}{Dw}
    \right \rangle_{\bar{L}}  \]
where $D$ is a differentiable operator involving partial derivatives of
order 1 to $k$ and $\bar{L} = L \times L^{n_k}$ with $n_k$ denoting the
number of multi-indices whose order is between 1 and $k$.

We recall the definition of the
Sobolev space $H^{k,p}(\Omega)$ as given in \cite{adams}.

\begin{definition}\rm
Suppose $\Omega$ is an open subset of $\mathbb{R}^n$.  $H^{k,p}(\Omega)$ is
the completion of $C^k(\Omega)$ with respect to the norm
$\|u\|_{k,p} = \Big( \sum_{|\alpha| \leq k} \|D^{\alpha} u\|_{L^p(\Omega)}^p
\Big)^{(1/p)}$, where $\alpha$ is a multi-index of order less than or
equal to $k$ and $D^{\alpha}$ is a partial derivative operator.
\end{definition}

It follows from this definition that $D$ is a closed and densely defined
operator on $L$.  By the assumption that $A$ is bounded on $L$ and hence
on $H$, $\| \cdot \|_{\alpha, H, A}$ induces an equivalent norm on $H$ for
each $\alpha$.  Hence the operator
%
\begin{equation*}
  T_{\alpha} = \binom{D}{\frac{1}{\sqrt{\alpha}} A}
\end{equation*}
%
is also a closed densely defined operator on $L$.  Thus there exists an
orthogonal projection $P_{\alpha}$ onto the graph of
$T_{\alpha} \subset \bar{L} \times K = L \times L^{n_k} \times K$.

From \eqref{e3} we have
%
\begin{equation*}
  \phi_{\alpha}'(u) h = \alpha \left\langle
  \binom {h}{T_{\alpha} h}, \binom {u}{T_{\alpha} u}
  \right \rangle_{\bar{L} \times K} - \left \langle
  \binom {h} {T_{\alpha} h }, \binom{0}{\binom{0}{\sqrt{\alpha} f}}
  \right \rangle_{\bar{L} \times K}.
\end{equation*}
%
Hence
%
\begin{align*}
  \phi_{\alpha}'(u) h
  &= \alpha \langle h,u \rangle_{\alpha, H ,A} -
    \left \langle P_{\alpha}\binom {h} {T_{\alpha} h }, \binom{0}
  {\binom{0}{\sqrt{\alpha} f}} \right \rangle_{\bar{L} \times K} \\
&=  \alpha \langle h,u \rangle_{\alpha, H, A} - \left \langle
    \binom {h} {T_{\alpha} h }, P_{\alpha}\binom{0}{\binom{0}{\sqrt{\alpha} f}}
    \right \rangle_{\bar{L} \times K}\\
& =  \alpha \langle h,u\rangle_{\alpha, H, A} - \left \langle h,\Pi P_{\alpha}
    \binom{0}{\binom{0}{\sqrt{\alpha} f}} \right \rangle_{\alpha, H, A},
\end{align*}
where $\Pi (x,y)=x$. Thus the gradient is
\[  
\nabla_{\alpha, H, A} \phi_{\alpha}(u) = \alpha u - \Pi P_{\alpha}\binom{0}
    {\binom{0}{\sqrt{\alpha} f}}. 
 \]
To obtain a more useful expression for the gradient, we
employ an expression for $P_{\alpha}$ due to von Neumann
(\cite{vonneumann}) and used for the development of the Sobolev gradient
theory in \cite{neuberger2010}:
%
\begin{equation*}
P_{\alpha} =
\begin{pmatrix}
  (I + T_{\alpha}^*T_{\alpha})^{-1}   &
  T_{\alpha}^*(I + T_{\alpha}T_{\alpha}^*)^{-1}  \\
  T_{\alpha}(I + T_{\alpha}^*T_{\alpha})^{-1}   &
  T_{\alpha}T_{\alpha}^*(I + T_{\alpha}T_{\alpha}^*)^{-1}
\end{pmatrix}
\end{equation*}
where $T_{\alpha}^*$ is the adjoint of $T_{\alpha}$ as a closed and densely
defined operator.
Using this expression, the final form of the gradient is
%
\begin{equation*}
  \nabla_{\alpha, H, A} \phi_{\alpha}(u) =  \alpha u -
  T_{\alpha}^*(I + T_{\alpha}T_{\alpha}^*)^{-1}\sqrt{\alpha} \binom{0}{f}.
\end{equation*}
%
In \cite{kazemi12} it is shown that
%
\begin{equation*}
  T_{\alpha}^*(I + T_{\alpha}T_{\alpha}^*)^{-1}\sqrt{\alpha} \binom{0}{f} =
  \alpha M A^* (\alpha I + A M A^*)^{-1} f
\end{equation*}
%
where $A^*$ is the adjoint of $A$ when viewed as a closed densely defined
operator on $L$, and $M$ is the embedding operator for $H$ and $L$; that is,
$\langle M x,y \rangle_H = \langle x,y \rangle_L$ for all $x \in L$ and
$y \in H$.  Thus
\[ \nabla_{\alpha, H, A} \phi_{\alpha}(u) =  \alpha u -
    \alpha M A^* (\alpha I + A M A^*)^{-1} f,  \]
and, the gradient is zero for
\begin{equation}\label{e6}
  u_{\alpha} =  M A^* (\alpha I + A M A^*)^{-1} f.
\end{equation}
Suppose that $f$ is in the range of $A$ and we seek a minimum-norm solution
$u \in H$ to \eqref{e1} as the limit of a sequence $u_{\alpha}$ as $\alpha$
approaches zero.  If the $L^2$ metric were used for regularization, the
limit of the sequence with convergence in the $L^2$ norm would not necessarily
have the required smoothness of an element of $H$.
With the Sobolev metric on the other hand, it follows from \eqref{e6}
that, for each $\alpha$, $u_{\alpha}$ is in the range of $M$ which is
a subspace of $H$.  Thus with convergence defined in the $H$ norm, the
limit of the sequence satisfies the desired regularity requirement.

We now make some remarks regarding properties of the operator
$MA^*(\alpha I + A M A^*)^{-1}$.

\begin{proposition}\label{properties}
Let $S_{\alpha} = ( \alpha I + A M A^*)^{-1}$.  Then the following
statements hold:
%
\begin{enumerate}
\item $S_{\alpha}$ is everywhere defined on $K$.
\item $S_{\alpha}$ is a bounded linear operator from $K$ to $K$ with norm
      less than or equal to $(1/\alpha)$.
\item $MA^*S_{\alpha}$ is a bounded linear operator from $K$ to $H$ with
      norm less than or equal to $(1/ \sqrt{\alpha})$.
\item $u_{\alpha} = MA^*(\alpha I + AMA^*)^{-1} f = (\alpha M^{-1} +
       A^*A)^{-1} A^*f$.
\end{enumerate}
\end{proposition}

\begin{proof}
Property 1 is proved in \cite[Theorem 3]{kazemi12}.  
To prove property 2,
simply note that $\langle S_{\alpha}^{-1} f,f \rangle_K \geq \alpha
\|f\|_K^2$ for all $f$ in the domain of $S_{\alpha}^{-1}$.  For case 3,
note that for a linear operator $J$ and $c$ a real number, 
$J(cI + J)^{-1} = I - c(cI + J)^{-1}$.  Thus 
$AMA^* S_{\alpha} = I - \alpha S_{\alpha}$.
Let $f \in K$.  Then
%
\begin{align*}
  \| MA^* S_{\alpha} f \|_H^2 
&= \langle MA^* S_{\alpha} f, MA^* S_{\alpha}
  f \rangle_H = \langle A^* S_{\alpha} f, MA^* S_{\alpha} f \rangle _L  \\
&=  \langle  S_{\alpha} f , AMA^* S_{\alpha} f \rangle_K=
  \langle  S_{\alpha} f , f - \alpha S_{\alpha} f \rangle_K \\
&\leq   \langle  S_{\alpha} f , f \rangle_K \leq (1 / \alpha) \|f\|_K^2.
\end{align*}
%
Thus $\| MA^* S_{\alpha} f \|_H \leq (1 / \sqrt{\alpha}) \|f\|_K$ and the
assertion follows.  Finally, to prove property 4, first note that since
$A$ is bounded on $L$ by assumption, $A^*$ is everywhere defined on $K$.
Thus
%
\begin{equation*}
  A^*f = A^* S_{\alpha}^{-1} S_{\alpha} f = (\alpha M^{-1} + A^* A)MA^*
  (\alpha I + AMA^*)^{-1}f.
\end{equation*}
%
Applying $(\alpha M^{-1} + A^*A)^{-1}$ to both sides of the equation gives
the desired result.
\end{proof}


\section{Convergence of the gradient flow} 

We consider now the gradient system
%
\begin{equation} \label{flowalpha}
z(0) = u_0 \in H \text{ and } z'(t)
= - \nabla_{\alpha, H , A} \phi_{\alpha} (z(t)).
\end{equation}
%
We restate \cite[theorems 4.1 and 7.1]{neuberger2010}.

\begin{theorem}
Suppose $F$ is a $C^1$ real-valued function defined on a Hilbert space $H$,
bounded below, and has a locally Lipschitzian derivative.  Then for any
initial state $u_0$, the gradient system \eqref{flowgen} has a unique global
solution.
\end{theorem}

\begin{theorem}
Suppose $F$ is a nonnegative $C^2$ function on a Hilbert space and is convex
in the sense of \ref{convexdef}.  Then the gradient system \eqref{flowgen}
has a unique global solution, and there exists $u \in H$ such that
%
\begin{equation*}
  u = \lim_{t \to \infty} z(t) \; \text{ and } \; \nabla_H F(u) = 0,
\end{equation*}
%
where convergence is in the strong sense using the $H$ norm.  Further, the
rate of convergence is exponential.
\end{theorem}

These two theorems establish that for any initial state $u_0$ the flow
\eqref{flowalpha} for $\phi_{\alpha}$ has a unique global solution and that
the flow converges strongly in the $H$ norm with an exponential convergence
rate.  Since $\phi_{\alpha}$ is convex, this minimum is unique. We denote
by $u_{\alpha}$ the unique minimizer of the energy obtained as the asymptotic
limit of the gradient system.  The expression in \eqref{e6} shows that
$u_{\alpha}$ is in the range of $M$ and hence in $H$.


\section{An example}  

Consider the problem of computing an approximation $u$ to the first derivative
$f'$ of a smooth function $f:[a,b] \to \mathbb{R}$ given only $f(a)$
and a set of $m$ discrete noise-contaminated data points $\{(x_i,y_i)\}$ with
\begin{equation*}
    a \leq x_1 \leq x_2 \leq \ldots \leq x_m \leq b,
\end{equation*}
and
\[  
y_i = f(x_i) + \eta_i, \quad (i = 1,\ldots,m)  
\]
for independent identically distributed zero-mean noise values $\eta_i$.
This problem is the inverse of the problem of computing
integrals and is ill-posed because the solution $f'$
does not depend continuously on the data $f$.  If we use divided
difference approximations to derivative values, then small
relative perturbations in the data can lead to arbitrarily large relative
changes in the solution, and the discretized problem is ill-conditioned.
We therefore require some form of regularization in order to avoid
overfitting.  Tikhonov regularization was first
applied to the numerical differentiation problem in \cite{cullum71}.

Let
\[  
\hat{f}(x) = f(x) - f(a)  
\]
for $x \in [a,b]$ so that $\hat{f}'(x) = f'(x)$ and $\hat{f}(a) = 0$.  
Then for $k = 0$, 1, or 2 and
$f \in H^{k+1}(a,b)$ the problem is to find $u \in H^k(a,b)$ such that
\begin{equation}
   A u(x) = \int_a^x u(t) \, dt = \hat{f}(x), \quad x \in [a,b].
\end{equation}
Note that $A$ is a bounded operator from $L^2(a,b)$ to $L^2(a,b)$ since
\[  
\|Au\|_{L^2}^2 = \int_a^b \left( \int_a^x u(t) \, dt\right)^2 \, dx
    \leq (b-a)^2 \int_a^b u^2.  
\]
We discretize the problem by partitioning the domain $[a,b]$ into $n$
subintervals of length $\Delta t = (b-a)/n$, and representing the
solution $u$ by the $n$-vector of midpoint values
\[ 
 u_j = u(t_j + \Delta t/2), \;\; (j = 1, \ldots, n)  
\]
for
\[  
t_j = a + (j-1) \Delta t, \;\; (j = 1, \ldots, n+1).  
\]
The discretized system is then
$A{\bf u} = {\bf \hat{y}}$, where $\hat{y}_i = y_i-f(a)$ and $A_{ij}$ is
the length of $[a,x_i] \cap [t_j,t_{j+1}]$:
\[ 
 A_{ij} = \begin{cases}
                       0        & \text{if } x_i \leq t_j  \\
                       x_i-t_j  & \text{if } t_j < x_i < t_{j+1}  \\
                       \Delta t & \text{if } t_{j+1} \leq x_i.
                     \end{cases}
\]
This linear system may be
underdetermined or overdetermined and is likely to be ill-conditioned.
We therefore use a least squares formulation with Tikhonov regularization.
We minimize the convex functional
\begin{equation}
   \phi({\bf u}) = \|A{\bf u}-{\bf \hat{y}}\|^2 + \alpha \|D{\bf u}\|^2,
\end{equation}
where $\alpha$ is a nonnegative regularization parameter, $\|\cdot\|$ denotes
the Euclidean norm, and $D$ is a differential operator of order $k$ defining
a discretization of the $H^k$ Sobolev norm:
\[  D^t = \begin{cases}
                    I                        & \text{if } k = 0  \\
                    (I \; D_1^t )            & \text{if } k = 1  \\
                    (I \; D_1^t \; D_2^t )   & \text{if } k = 2,
                  \end{cases} 
 \]
where $I$ denotes the identity matrix, and $D_1$ and $D_2$ are first
and second difference operators.  We use second-order central differencing
so that $D_1$ maps midpoint values to interior grid points, and $D_2$ maps
midpoint values to interior midpoint values.  The regularization
(smoothing) parameter $\alpha$ defines a balance between fidelity to the
data on the one hand, and the size of the solution norm on the other
hand.  The optimal value depends on the choice of norm.  Larger values of $k$
enforce more smoothness on the solution.  Setting the gradient of $\phi$ to
zero, we obtain a linear system with an order-$n$ symmetric positive definite
matrix:
\[  
(A^t A + \alpha D^t D){\bf u} = A^t {\bf \hat{y}}.  
\]
In the case that the error norm $\|\eta\|$ is known, a good value of
$\alpha$ is obtained by choosing it so that the residual norm
$\|A{\bf u}-{\bf \hat{y}}\|$ agrees with $\|\eta\|$.  This is Morozov's
discrepancy principle (\cite{morozov67}).


\begin{figure}[htbp]
\begin{center}
\includegraphics[width=0.7\textwidth]{fig1} %f0.png
\end{center}
  \caption{Computed approximation to $f', k=0, f(x) = \cos(x)$}
  \label{f1}
\end{figure}

\begin{figure}[htbp]
\begin{center}
\includegraphics[width=0.7\textwidth]{fig2} %f1.png
\end{center}
    \caption{Computed approximation to $f', k=1, f(x) = \cos(x)$}
    \label{f2}
\end{figure}

\begin{figure}[htbp]
\begin{center}
\includegraphics[width=0.7\textwidth]{fig3} %f2.png
\end{center}
  \caption{Computed approximation to $f', k=2, f(x) = \cos(x)$}
    \label{f3}
\end{figure}

We chose the test function $f(x) = \cos(x)$ on $[-1,1]$ and created
a data set consisting of $m = 500$ points with uniformly distributed
abscissae $x_i$ and data values $y_i = f(x_i) + \eta_i$, where $\eta_i$
is taken from a normal distribution with mean 0 and standard deviation
$\sigma = 0.1$.  We used the known value of $\|\eta\|$ to compute the
optimal parameter value $\alpha$.  The maximum relative error in the
computed derivative decreased rapidly as $k$ increased:  $1.001, 0.2640$,
and $0.0749$ corresponding to $k = 0, k = 1,$ and $k = 2$, respectively.
The solutions are graphed in Figure \ref{f1}, \ref{f2}, and \ref{f3}.


\begin{thebibliography}{99}

\bibitem{adams} R. A. Adams, J. F. Fournier;
 Sobolev Spaces, 2nd Edition,
   Academic Press, 2003.

\bibitem{aster13} R. C. Aster, B. Borchers, C. H. Thurber;
 Parameter
   Estimation and Inverse Problems, second edition, Academic Press, 2013.

\bibitem{cullum71}  J. Cullum;
 Numerical differentiation and  regularization, SIAM J. Numer. Anal. 8 (1971), 254-–265.

\bibitem{enghanke} H. W. Engl, M. Hanke,  A. Neubauer;
 Regularization of
   Inverse Problems. Mathematics and Its Applications. Kluwer Academic
   Publishers, Dordrecht, 1996.

\bibitem{hadamard02} J. Hadamard;
 Sur les probl\`emes aux d\'eriv\'ees partielles et
   leur signification physique (1902), 49--52.

\bibitem{hofmann09}  B. Hofmann, P. Math\'{e}, H. von Weizs\"{a}cker;
   Regularization in Hilbert space under unbounded operators and general
   source condition, Inverse Problems, 25(11):115013 (15pp), 2009.

\bibitem{kabanikhin08}  S. I. Kabanikhin;
 Definitions and examples   of inverse and ill-posed problems, 
J. Inv. Ill-Posed Problems 16 (2008),    317--357.

\bibitem{kazemi12} P. Kazemi, R. J. Renka;
 A Levenberg-Marquardt method
   based on Sobolev gradients, Nonlinear Analysis 75 (2012) 6170--6179.

\bibitem{morozov67}  V. A. Morozov;
 Choice of a parameter for  the solution of functional equations by the 
regularization method,
  Sov. Math. Doklady 8 (1967), 1000--1003.

\bibitem{neuberger2010} Neuberger J. W.;
 Sobolev Gradients and Differential Equations, 2nd edition, Springer, 2010.

\bibitem{ramm08} A. G. Ramm;
 On unbounded operators and applications, Appl.
   Math. Lett. 21 (2008), 377--382.

\bibitem{riesznagy} Riesz, Sz. Nagy;
 Functional Analysis, Dover  Publications, Inc., 1990.

\bibitem{tikhonov63}  A. N. Tikhonov;
 Regularization of  incorrectly posed problems, 
Sov. Math. Dokl. 4 (1963), 1624-–1627.

\bibitem{tikhonov77}  A. E. Tikhonov, V. Y. Arsenin;
  Solutions  of Ill-posed Problems, John Wiley \& Sons, New York, 1977.

\bibitem{tikhonov98} A. N. Tikhonov, A. S. Leonov, A. G. Yagola;
 Nonlinear  Ill-posed Problems. Number 14 in Applied Mathematics
 and Mathematical  Computation. Chapman and Hall, London, 1998.

\bibitem{vonneumann} J. von Neumann;
 Functional Operators II, Annls. Math.    Stud., 22, 1940.

\end{thebibliography}

\end{document}
