next up previous contents
Next: Sequential interpretation Up: Choice of the nudging Previous: Choice of the nudging   Contents

Variational interpretation of the nudging

The standard nudging method has been widely studied in the past decades [69,109,103,102]. Thus, there are several ways to choose the nudging matrix $ K$ in the forward part of the algorithm. One can for example consider the optimal nudging matrix $ K_{opt}$ , as discussed in [119,112]. In such an approach, a variational data assimilation scheme is used in a parameter estimation mode to determine the optimal nudging coefficients. This choice theoretically provides the best results for the forward part of the BFN scheme, but the computation of the optimal gain matrix is costly.

When $ K=0$ , the forward nudging problem (3.2) simply becomes the direct model (3.1). On the other hand, setting $ K=+\infty$ forces the state variables to be equal to the observations at discrete times, as is done in [105,104]. These two choices have the common drawback of considering only one of the two sources of information (model and data).

Let us assume that we know the statistics of errors on observations, and denote by $ R$ the covariance matrix of observation errors. This matrix is involved in all standard data assimilation, either variational (3D-VAR, 4D-VAR, 4D-PSAS, ...) or sequential (Kalman filters) [52,59,90,65,66]. Usually, it is impossible to know the exact statistics of errors, and thus only an approximation of $ R$ is available, assumed to be symmetric positive definite.

We assume here the direct model to be linear (or linearized). We consider a temporal discretization of the forward nudging problem (3.2), using for example an implicit scheme. If we denote by $ X^n$ the solution at time $ t_n$ and $ X^{n+1}$ the solution at time $ t_{n+1}$ , and $ \Delta t=t_{n+1}-t_n$ , then equation (3.2) becomes

$\displaystyle \frac{X^{n+1}-X^n}{\Delta t} = F X^{n+1} + K(X_{obs} - C X^{n+1}).$ (3.7)

We now set the nudging matrix to be

$\displaystyle K = C^T R^{-1}.$ (3.8)

Then, it is straightforward to see that problem (3.7) is equivalent to the following optimization problem:
$\displaystyle X^{n+1}=\textrm{arg}\min_X$   $\displaystyle \hspace*{-0.5cm}\left[ \displaystyle\frac{1}{2} \langle X-X^n,X-X^n\rangle - \frac{\Delta t}{2} \langle FX,X\rangle \right.$ (3.9)
    $\displaystyle \hspace*{-0.5cm} + \left. \frac{\Delta t}{2} \langle R^{-1}(X_{obs}-CX),X_{obs}-CX\rangle \right].$  

The first two terms correspond exactly to the energy of the discretized direct model, and the last term is the observation part of the variational cost function. This variational principle shows that at each time step, the nudging state is a compromise between minimizing the energy of the system and the distance to the observations.

As a consequence, there is no need to consider an additional term ensuring an initial condition close to the background state like in variational algorithms, neither for stabilizing or regularizing the problem, nor from a physical point of view. One can simply initialize the BFN scheme with the background state, without any information on its statistics of errors.

The nudging method naturally provides a correction to the model equations from the observations. The model equations are hence weak constraints in the BFN scheme. In some nonlinear cases, the $ \langle FX,X\rangle$ term in equation (3.9) can be replaced by $ -G(X)$ , where $ G$ is the energy of the system at equilibrium.


next up previous contents
Next: Sequential interpretation Up: Choice of the nudging Previous: Choice of the nudging   Contents
Back to home page