E cient Numerical Algorithms for Solving Nonlinear Filtering Problems by Feng Bao A dissertation submitted to the Graduate Faculty of Auburn University in partial ful llment of the requirements for the Degree of Doctor of Philosophy Auburn, Alabama August 2, 2014 Keywords: Nonlinear lter, Bayesian theory, implicit lter, Zakai lter, sparse grid, backward doubly stochastic di erential equation, It^o-Taylor expansion Copyright 2014 by Feng Bao Approved by Yanzhao Cao, Chair, Professor of Mathematics Amnon Meir, Co-Chair, Professor of Mathematics Paul Schimdt, Professor of Mathematics Erkan Nane, Professor of Mathematics Jianjun Dong, Professor of Physics Abstract Examples of nonlinear ltering problems arise in biology, mathematical nance, signal processing, image processing, target tracking and many engineering applications. Commonly used numerical simulation methods are the Bayesian lter which is derived from the Bayesian formula and the Zakai lter which is related to a system of stochastic partial di erential equation known as \the Zakai equation". This dissertation mainly focuses on developing and analysing novel, e cient numerical algorithms for solving nonlinear ltering problems. We rst introduce a novel numerical algorithm which lies in the general framework on the Bayesian lter. The algorithm is constructed based on samples of the current state obtained by solving the state equation implicitly. We call this algorithm the \implicit lter method". Rigorous analysis has been done to prove the convergence of the algorithm. Through numerical experiments we show that our algorithm is more accurate than the Kalman lter and more stable than the particle lter. In the second e ort of this work, we propose a hybrid numerical algorithm for the Zakai lter to solve nonlinear ltering problems e ciently. The algorithm combines the splitting- up nite di erence scheme and hierarchical sparse grid method to solve moderately high dimensional nonlinear ltering problems. When applying hierarchical sparse grid method to approximate bell-shaped solutions in most applications of nonlinear ltering problem, we introduce a logarithmic approximation to reduce the approximation errors. Some space adaptive methods are also introduced to make the algorithm more e cient. Numerical experiments are carried out to demonstrate the performance and e ciency of our algorithm. ii In this dissertation, we also develop high order numerical approximation methods for backward doubly stochastic di erential equations (BDSDEs). One of the most importan- t properties of BDSDEs is it?s equivalence to the Zakai equation. In this connection, our numerical approximation methods for BDSDEs can be considered as e cient numerical ap- proaches to solving nonlinear ltering problems. The convergence order is proved through rigorous error analysis for each algorithm. Numerical experiments are carried out to verify the theoretical results and to demonstrate the e ciency of the proposed numerical scheme. iii Acknowledgments This dissertation would never have been possible without the guidance of my committee memebers, help from friends, and support from my family and wife. Foremost, I am heartily thankful to my advisors, Dr. Yanzhao Cao and Dr. Amnon J. Meir. Through my ve-year study at Auburn University, they provided excellent teaching, great company, and a signi cant amount of new ideas. I must also thank Dr. Yanzhao Cao who encouraged me to step in the deep ocean of the stochastic lter, for his unconditional support and the full con dence he always o ered. I would like to thank my committee members for their helpful comments to my disserta- tion. I would also like to thank all the collaborators who always willing to help and give their best suggestions. Besides, I am very grateful to Dr. Weidong Zhao who gives great ideas in my research of numerical methods for backward doubly stochastic di erential equations. Finally, I would like to express my deepest grateful to my beloved wife Ye and my parents for their constant support. Without them, this dissertation would have hardly been possible. iv Table of Contents Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ii Acknowledgments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv List of Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii List of Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . x 1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 2 Nonlinear Filtering problems and Existing Numerical Approaches . . . . . . . . 6 2.1 Mathematical Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 2.2 Nonlinear Filtering Problems . . . . . . . . . . . . . . . . . . . . . . . . . . 9 2.3 Bayesian Filters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 2.3.1 Extended Kalman Filter . . . . . . . . . . . . . . . . . . . . . . . . . 11 2.3.2 Particle lter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 2.4 Zakai Filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 3 An Implicit Algorithm of Solving Nonlinear Filtering Problems . . . . . . . . . . 15 3.1 Methodology and the implicit lter algorithm . . . . . . . . . . . . . . . . . 15 3.1.1 Bayesian Optimal Filter . . . . . . . . . . . . . . . . . . . . . . . . . 16 3.1.2 An inverse algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 3.2 Weak Convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 3.3 Numerical experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 4 A Hybrid Sparse Grid Approach for Nonlinear Filtering Problems Based on Adaptive-Domain of the Zakai Equation Approximations . . . . . . . . . . . . . 38 4.1 Nonlinear Filtering Problems and Zakai Equations . . . . . . . . . . . . . . . 38 4.2 Hierarchical Local Sparse Grid Interpolation . . . . . . . . . . . . . . . . . . 40 4.2.1 Standard hierarchical sparse grid ipnterpolation . . . . . . . . . . . . 40 v 4.2.2 Hierarchical sparse grid approximation of bell-shaped cpurves . . . . 43 4.3 Hybrid Approach for Numerical Solution of the Nonlinear Filtering Problem. 45 4.3.1 Adaptive selection of solution domains . . . . . . . . . . . . . . . . . 46 4.3.2 Spliting-up nite di erence method on sparse grid . . . . . . . . . . . 48 4.4 Numerical Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51 5 Numerical Algorithms for Backward Doubly Stochastic Di erential Equations and it?s Applications to Nonlinear Filtering Problems . . . . . . . . . . . . . . . 63 5.1 FBDSDEs and SPDEs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 5.2 Half Order Numerical Algorithms . . . . . . . . . . . . . . . . . . . . . . . . 66 5.2.1 Numerical algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . 68 5.2.2 Error Estimates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 5.2.3 Numerical experiments . . . . . . . . . . . . . . . . . . . . . . . . . . 86 5.3 First Order Numerical Algorithms . . . . . . . . . . . . . . . . . . . . . . . . 89 5.3.1 Variational Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 5.3.2 Reference equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 5.3.3 Numerical scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100 5.3.4 Error Estimates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 5.3.5 Fully discrete scheme and its error estimate . . . . . . . . . . . . . . 113 5.3.6 Numerical experiment . . . . . . . . . . . . . . . . . . . . . . . . . . 124 6 Concluding Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129 Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131 vi List of Figures 3.1 Original Position . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 3.2 Extended Kalman Filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 3.3 Particle Filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 3.4 implicit Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 3.5 Target Positions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 3.6 Comparison result for the observation angle . . . . . . . . . . . . . . . . . . . . 36 3.7 Simulation result of EKF . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 3.8 Comparison result between Particle Filter and implicit Algorithm . . . . . . . . 37 4.1 The interpolation error for regular sparse grid approximation up to 5 dimensions. 44 4.2 The interpolation error for logarithmic sparse grid approximation up to 5 dimen- sions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 4.3 (a) is the x1x2-marginal surface of U with x3 = x4 = 0 while (b) is its interpolate approximation obtained by applying the regular sparse grid approximation to function U. We can see from the Figure that the approximation is quite poor and signi cant oscillations occur at the bottom of the surface. On the other hand, from (d) one can see that the logarithmic interpolation approximation described above is far more accurate. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 vii 4.4 Target Trajectory and Adaptive Solution Domain. The red curve shows the real target state. The blue points are actual states of the target and the blue boxes are the corresponding solution domains. . . . . . . . . . . . . . . . . . . . . . . 52 4.5 Target state PDF at time step: (a) n=1; (b) n=50; (c)=80; in the corresponding solution domain . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53 4.6 2D nonlinear ltering problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 4.7 Marginal distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 4.8 Sparse grid Zakai lter estimate of posterior mean with 95% probability region . (a) Estimate of posterior mean and 95% probability region: X-coordinate (b) Estimate of posterior mean and 95% probability region: Y-coordinate . . . . . . 57 4.9 Bearing-only tracking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58 4.10 Comparison of estimated mean of the target state between particle lter and sparse grid Zakai lter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 4.11 Sparse grid Zakai lter estimate of posterior mean with 95% probability region. (a) Estimate of posterior mean of the location x and 95% probability region. (b) Estimate of posterior mean of the location y and 95% probability region. (c) Estimate of posterior mean of the velocity u and 95% probability region. (d) Estimate of posterior mean of the velocity v and 95% probability region . . . . 60 4.12 Marginal distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 5.1 Example 1: Convergence comparison between the direct nite di erence scheme and our scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 5.2 Example 2: Convergence comparison between the approximations of y and z . . 89 viii 5.3 Target tracking by using one detector . . . . . . . . . . . . . . . . . . . . . . . . 126 5.4 Tracking Estimate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126 5.5 Probability Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127 ix List of Tables 4.1 Comparison of computing costs . . . . . . . . . . . . . . . . . . . . . . . . . . . 57 4.2 Comparison of computing costs . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 5.1 Example 1 of Section 6.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87 5.2 Example 2 of Section 6.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 5.3 Example 1 of Section 6.3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125 x Chapter 1 Introduction Stochastic ltering theory was rst established in the early 1940s due to the pioneering work by Norbert Wiener [79, 80] and Andrey N. Kolmogorov [46]. The goal of data assimila- tion of the stochastic ltering problem is to obtain good estimates of the state of a stochastic dynamic system model based on noisy partial observations. Examples of stochastic ltering problems arise in biology [4], mathematical nance [11, 31, 37, 69], signal processing [2, 3, 74], image processing [12, 71, 91], target tracking [13, 20, 26, 29, 62, 85] and many engineering applications. The major breakthrough of the classic ltering problem was the landmark work of Kalman and Bucy in 1960s [45], and subsequent Kalman-Bucy lter in 1961. Most people now call this type of lter theory \Kalman lter"( see also [28, 40, 49, 56, 57, 66, 72, 77] ). For decades, Kalman lter has been one of the dominant tools in solving ltering prob- lems. According to the Kalman lter, under standard assumption of linearity and Gaussinity of the noise, the conditional distribution of the state, given the observations, is Gaussian. This conditional distribution gives the best estimate of the statistical description of the state of the system based on all the available observation information up to the current time. The power of Kalman lter lies in it?s simplicity and accuracy. Largely because of the success of Kalman lters, linear and nonlinear lters have been applied in the various engineering and scienti c areas, including communications such as positioning in wireless networks, signal processing such as tracking and navigation, economics and business, and many others. In most of the practical application problems, however, linearity assumption is not valid because of the nonlinearity in the model speci cation process as well as the observation process. For example, in typical atmospheric data assimilation problems, the dynamic system 1 is described by a rather complex system of equations, which is highly nonlinear. In bearing- only tracking problems, depending on how the measurements are obtained, the observation could also have a nonlinear relation with the state process. Numerous nonlinear ltering methods have been proposed and developed to overcome the linearity limitation of Kalman lter. Two of the widely used methods for nonlinear ltering problems are the Bayesian lter and Zakai lter. The Bayesian lter is based on the Baysian theory. It is originally discovered by the British researcher Thomas Bayes in 1763. The most well-known Bayesian lter methods include the extended Kalman lter (EKF) and Particle Filter method (PFM). For the EKF, the state equation and the observation equation are linearized so that the standard Kalman lter can be applied [7, 14, 27, 33, 42, 43, 44, 47, 63, 73]. The central theme behind the PFM involves the representation of the desired probability density function (PDF) of the system state with a set of adaptively selected random samples [15, 18, 21, 24, 38, 59, 58, 65, 76, 78]. Since PFMs are essentially sequential Monte Carlo methods, with su ciently large number of samples, PFMs can provide an accurate representation of the PDF for the nonlinear ltering solution. While the aforementioned methods have been remarkably successful in attacking the nonlinear ltering problem, each of them has its drawbacks and limitations. For instance, when the state equation describing the signal process and the observation equation are highly non-linear, the EKF can give particularly poor performance. PFM has a number of advan- tages over EFK, including its ability to represent arbitrary densities, adaptively focus on the most probable regions of state-space. However, it also has a number of disadvantages, including high computational complexity, degeneracy for long period simulation and its di - culty of determining optimal number of particles. The rst e ort of this work is to construct a new algorithm for numerical simulations of nonlinear ltering problems using the Bayesian ltering theory. For each time recursive step, we have two stages: the prediction state and the update stage. The prediction stage gives the estimation for the prior PDF of the future state based on the currently available observation information while update stage gives the 2 posterior PDF from the updated observation information and the the result obtained in the prediction stage. However, instead of attempting to search for a representation of the PDF as in PFM, we approximate the PDF as a function over a grid in state space. Speci cally, at the prediction stage, we attempt to seek the predicted pdf of the future state variable through a Monte Carlo method by evaluating the conditional expectation of the future state with respect to the current stage. Since the sample points for the current state is computed by solving the state equation implicitly, we name our method as an \implicit lter method". The following two items summarize the novelty of our approach. (i) We propose an accurate implicit scheme for prediction purpose. The implicit scheme has stabilizing e ect on the proposed numerical algorithm. This is veri ed in our numerical experiments. (ii) Based on the Bayesian theory, we apply a novel Monte-Carlo like method to approxi- mate the conditional expectation in the update stage to compute the prior PDF. The Zakai lter represents the PDF of the nonlinear ltering solution through the solu- tion of a parabolic-type stochastic partial di erential equation, known as the Zakai equation [82]. Similar to the PFM, the Zakai lter allows one to accurately compute conditional dis- tributions. A number of numerical algorithms have been proposed to solve Zakai equations [1, 10, 22, 23, 30, 32, 36, 39, 67, 75, 85]. One of the most e ective methods is the splitting-up approximation scheme [10, 30, 85] where the original Zakai equation is split into a second order deterministic PDE, related with the prediction step, and a degenerated second order stochastic PDE in the update step. In the numerical simulation process, a prior PDF is obtained by solving the deterministic PDE at the prediction; then this prior PDF is updated following a posteriori criterion. The main drawback of the Zakai lter is that the numerical approximation is grid based, thus it su ers the so-called \curse of dimensionality" since the computing cost increases exponentially as the dimension of the system increases. Another di culty related to solutions of the Zakai equation is that the domain is the whole space 3 Rd. To address these challenges, we propose the construction of an e cient hybrid numer- ical algorithm which combines the advantages of the splitting-up approximation scheme for the Zakai equation, a hierarchical sparse grid method [41, 50, 55, 68, 70, 81, 83, 84] for moderately high dimensional nonlinear ltering problems to compute the numerical solution of the Zakai lter, and an importance sampling method to adaptively construct a bounded domain at each time step of the temporal discretization. Speci cally, this enables us to use the splitting-up nite di erence scheme to solve the Zakai equation on the sparse grid of the bounded domain. The hierarchical sparse grid method, which was originally created to approximate multi-variable functions, uses only O(n(logn)d 1) number of grid points instead of O(nd) number of grid points, required by the standard full-grid approximation. Backward doubly stochastic di erential Equations (BDSDEs) were introduced as Feynman- Kac type probabilistic representations of semi-linear parabolic stochastic partial di erential equations (SPDEs), which are generalized Zakai equations [61]. In this connection, numerical approximation methods for simulating solutions of BDSDEs are also numerical approxima- tion methods for solving Zakai type equations. Thus it can be considered as alternative numerical approaches for nonlinear ltering problems. In this work, we propose e cient numerical algorithms for approximating solutions of BDSDEs. Several e ective numerical approaches for backward stochastic di erential equations ( BSDEs ) and forward backward stochastic di erential equations ( FBSDEs ) have been proposed in the last decade, including primary schemes for BSDEs ([5, 19, 52, 86]), the four-step scheme for FBSDEs [9, 53, 54, 51]), and the -scheme with high convergence rate for BSDEs ([87, 88, 89, 90]). In comparison, e cient high order numerical algorithms for BDSDEs are not well developed. Obviously solving BDSDEs numerically is more di cult than solving SDEs and BSDEs as they contain two Brownian motions. In this work, we rst construct a half order algorithm using the simple Euler method. It is much more involved to construct higher order algorithms. The bottle neck is the di culty in approximating the forward and backward It^o integrals with high order quadratures. To tackle this di culty, we propose to use an It^o-Taylor formula 4 for two-sided stochastic integrals [60, 61] in order to obtain a higher order quadrature for the backward It^o integral; for the forward stochastic integral, we propose to use the varia- tional equations for BDSDEs [61] to derive high order quadrature rule. It is worth noting that although our focus is the on construction of a rst order algorithm, the methodology developed in this work can be used to obtain even higher order algorithms. The outline of this work is as follows. In Chapter 2, we give a brief overview of the prob- ability theory that will be used throughout the rest of the chapters. We will also introduce the mathematical de nition of the nonlinear ltering problems and some existing numerical approximation methods for nonlinear ltering problems. In Chapter 3, we present our novel \implicit lter algorithm" for nonlinear ltering problems. We also give convergence analysis which shows weak convergence of our algorithm. Numerical experiments demonstrate that our algorithm is more accurate than the Kalman lter and more stable than the particle l- ter. In Chapter 4, we construct a hybrid nite di erence algorithm for the Zakai equation to solve nonlinear ltering problems. The algorithm combines the splitting-up nite di erence scheme and hierarchical sparse grid method to solve moderately high dimensional nonlinear ltering problems. A space adaptive method is introduced to make the algorithm more e - cient. Numerical experiments are carried out to demonstrate the performance and e ciency of our algorithm. In Chapter 5, we focus on numerical algorithms of BDSDEs, which can be considered as an alternative numerical approach for nonlinear ltering problems. We rst introduce a half order convergence algorithm based on Euler approximation for stochastic integrals. Then, we develop a rst order convergence scheme by using two-sided It^o Taylor expansion. For each algorithm, we give rigorous error analysis and numerical experiments to verify the convergence rates of our algorithms. 5 Chapter 2 Nonlinear Filtering problems and Existing Numerical Approaches In this Chapter, we rst give some de nitions from general probability theory. Then, we introduce the mathematical formulations of nonlinear ltering problems and some of the most well-known numerical schemes that have been studied in the literature. There are mainly two numerical approaches to solve the nonlinear ltering problem. One is the Bayesian lter, which is based on the Bayes rule; the other is the Zakai lter, which is based on the Zakai equation, a stochastic evolution equation. 2.1 Mathematical Preliminaries De nition 1 If is a given set, then a -algebra F on is a family F of subsets of with the following properties: (i) ;2F (ii) F 2F)FC 2F, where FC = nF is complement of F in (iii) A1;A2; 2F)A := S1i=1Ai2F. The pair ( ;F) is called a measurable space. A probability measure P on a measurable space ( ;F) is a function P :F![0;1] such that (a) P(;) = 0, P( ) = 1 (b) if A1;A2; 2F and fAig1i=1 is disjoint (i. e. Ai\Aj =; if i6= j) then P 1[ i=1 Ai ! = 1X i=1 P(Ai): 6 The triple ( ;F;P) is called a probability space.p Given any family U of subsets of there is a smallest -algebra HU containing U, namely HU =\fH;H -algebra of ; U Hg: We call HU the -algebra generated by U. If ( ;F;P) is a given probability space, then a function f : ! Rn is called F- measurable if f 1(U) :=f!2 ; f(!)2Ug2F for all open sets U2Rn. If X : !Rn is any function, then the -algebra HX generated by X is the smallest -algebra on containing all sets X 1(U); U Rn open: De nition 2 A random variable X is an F-measurable function X : ! Rn. Every random variable induces a probability measure X on Rn de ned by X(B) = P(X 1(B)): X is called the distribution of X. If R jX(!)jdP(!) <1 then the number E[X] := Z X(!)dP(!) = Z Rn xd X(x) is called the expectation of X (w.r.t. P). De nition 3 A stochastic process is a parameterized collection of random variables fXtgt2T de ned on a probability space ( ;F;P) and assuming values in Rn. 7 Next, we introduce a general form of the stochastic di erential equations. First, we need the following de nition of Brownian motion De nition 4 Brownian motion (or Wienner process) Bt starting at x is a stochastic process satisfying the following properties (i) B0 = x (ii) Bt is almost surely continuous (iii) Bt has independent increments (iv) Bt Bs N(0;t s) (for 0 s t ) N( ; 2) denotes the normal distribution with mean value and variance 2. To proceed, we show how to de ne the It^o integral RTS f(t;!)dBt(!). A function is called elementary if it has the form (t;!) = X j ej(!) X[tj;tj+1)(t): For elementary function , we de ne the It^o integral by Z T S (t;!)dBt(!) = X j 0 ej(!)[Btj+1 Btj](!): Then we de ne the It^o integral as follows De nition 5 Let f be a measurable function. Then the It^o integral of f (from S to T) is de ned by Z T S f(t;!)dBt(!) = limn!1 Z T S n(t;!)dBt(!) where f ng is a sequence of elementary functions such that E[ Z T S (f(t;!) n(t;!))2dt]!0 as n!1: 8 With the de nition of It^o integral, a general form of stochastic di erential equation is given by dXt = b(Xt)dt+ (Xt)dBt; X0 = x; where Xt is the solution of the stochastic di erential equation. 2.2 Nonlinear Filtering Problems Now, let us consider the following generic stochastic ltering problem in a dynamic state-space form Xt dt = f(t;Xt; _Wt) (2.1) Yt dt = g(t;Xt; _Bt): (2.2) Equations (2.1) and (2.2) are called signal state equation and measurement equation (or observation equation), respectively; Xt represents the state vector, Yt is the measurement vector; f : Rnx ! Rnx and g : RNx ! RNy are two vector valued functions, which are potentially time-varying; Wt and Bt are two independent Wienner processes, with covariance I and R, respectively, which represent process noise and measurement noise, respectively. When f and g are both linear functions, the ltering problem (2.1)-(2.2) is called the linear ltering problem, otherwise, it?s called the nonlinear ltering problem. In many applications, the noise can be assumed to be additive and (2.1)-(2.2) become dXt = b(Xt)dt+ (Xt)dWt (2.3) dYt = h(Xt)dt+dBt; (2.4) where, b : Rnx ! Rnx and h : RNx ! RNy are two vector valued functions, which are potentially time-varying, and : Rnx !Rnx nw is a matrix valued function. 9 The main purpose of numerical simulations of a ltering process is to obtain, recursively in time, the best estimate for the probability density function (pdf) of the state Xt based on observations up until time t, i.e.fYs : 0 s tg. This can be expressed as nding stochastic process ~Xt such that ~Xt = E[(XtjYt)] = inffE[jX Yj2];Y 2Kg where Yt is the -algebra generated by the observation process up to t, and K is the space of all Yt-measurable and square integrable random variables. In typical applications, observations are available only at a discrete sequence of time instants. We assume that the observations are collected at time instants t;2 t; , where t is a given time stepsize. Then denoting Yn := Y(n t) Y((n 1) t); we have Yn = Z n t (n 1) t h(Xs)ds+Bn t B(n 1) t h(Xn t) t+ n; where n is an i.i.d. sequence of standard Gaussian variables and := pR t. Denoting h(x) t by g(x), we rewrite a discrete analog of the observation equation (2.4) as follows. Yn = g(Xn t) + n; n 1: (2.5) Alternatively, one can consider (2.5) as the original underlying observation model and not a discrete approximation to (2.4). 10 2.3 Bayesian Filters We rst review the extended Kalman lter (EKF), the most commonly used numerical approach for solving the nonlinear lter. Then, we discuss a popular class of approximation methods, the particle lter. 2.3.1 Extended Kalman Filter The basic idea of EKF is to linearize the nonlinear functions in both the state equation and observation equation and as a result approximate the conditional distribution Pn by a normal distribution with parameters ( n; n). Given the approximation P n 1 to Pn 1 as a normal distribution with mean n 1 and covariance n 1. The approximation P n to Pn is obtained by the following two steps. Prediction step. Let fx (t);t 0g be the trajectory about which linearization is performed. Typically, the trajectory x is chosen as the solution of the deterministic ODE _xt = b(xt) with initial condition n 1. We denote Xt := Xt x (t). Then, the linearized dynamic signal state is given as _Xt = [@f@x]x=x Xt + (x (t))dBt: The prediction step of the standard continuous time Kalman lter is then applied with this linearized state model to give the approximation ^P n to ^Pn via a normal distribution with parameters (^ n;^ n). Update step. Similar to the state model, the observation model (2.5) is also linearized and is given as follows. Yn g(^ n) = [@g@x]x= ^mn(Xtn ^mn) + n 11 The update step of the usual Kalman lter [45] is used with this linearized observation model to obtain P n as a normal distribution with parameter n; n. 2.3.2 Particle lter The particle lter, also known as sequential Monte-Carlo method is a recursive Bayesian lter based on the idea of approximating expected values by suitable Monte-Carlo sample averages. It is also called bootstrap lter, due to Gordon, Salmond and Smith [38]. The basic idea of the particle lter is as follows. The state space is partitioned into subregions, in which some random samples, called particles, are lled according to some probability measure. The higher the probability of a subregion, the denser the particles are concentrated. Since the pdf can be approximated by the point-mass histogram, by random sampling of the state space, we get a number of particles representing the pdf. To be more speci c, the approximation P n to Pn is given by a discrete probability measure supported on random points x(n)1 ; ;x(n)L and corresponding weights p(n)1 ; ;p(n)L . Here, fx(n)l gl=1; ;L are called particles and L represents the number of particles that are used to approximate the distribution Pn. Given the approximation P n to Pn by the discrete probability measuref(x(n 1)1 ;p((n 1))1 ); (x(n 1)L ;p((n 1))L )gthe two key steps of the algorithm are as follows. Prediction step. Propagate each of the particles according to the state equation, i.e. x(n 1)j ! ^x(n)j . This requires simulating the SDE (2.3) with initial condition x(n 1)j by some discretization scheme. For example, we can use Euler scheme to get ^x(n)j = x(n 1)j +b(x(n 1)j ) t+ (x(n 1)j )p t n 1; wheref ngn=0;1;2;::: is an i.i.d sequence of standard Gaussian random variables. This gives an approximation ^P n to ^Pn as the discrete probability distribution f(^x(n)1 ; ^p(n)1 ); (^x(n)L ; ^p(n)L )g where we set ^p(n)j := p(n 1)j . 12 Update step. Update the weights ^p(n)j !p(n)j using the observation Yn by setting p(n)j = c^p(n)j R(^x(n)j ;Yn); where c is a normalization constant and R(^x(n)j ;Yn) := exp(g(^x (n) j )Yn 1 2jg(^x (n) j )j 2 2 ): The approximation P n is now given as f(x(n)1 ;p(n)1 ); (x(n)L ;p(n)L )g where we set x(n)j := ^x(n)j . Although the scheme is very easy to implement, it su ers from several degeneracy problems, especially in high dimensions. The main di culty is that after a few time steps all the weights tend to concentrate on a very few particles which drastically reduces the e ective sample size. A common remedy for this problem is to re-sample all the particles in order to rejuvenate the particle cloud. 2.4 Zakai Filter An alternative approach to the computation of the nonlinear lter is by developing an evolution equation for the conditional pdf p(t;Xt) := p(XtjYt) (see [82]). To be more speci c, under suitable regularity conditions one can show that the conditional pdf p(t;x) is the unique solution of the following stochastic partial di erential equation (SPDE): dp(t;x) = L p(t;x)dt+h(x)p(t;x)dYt; (2.6) where L is the in nitesimal generator of the state equation Xt and L is the adjoint of L. The above SPDE is also called the Zakai equation. The Zakai equation gives a recursive way for evaluating p(t;x). Several works [10, 23, 30, 36, 85] have been done to develop suitable time and space discretization schemes for the Zakai equation for the purpose of obtaining 13 numerical approximations for the nonlinear lter. This type of approximation methods are called \the Zakai lters". 14 Chapter 3 An Implicit Algorithm of Solving Nonlinear Filtering Problems In this Chapter, we consider the following state and observation equations in the dy- namic state-space form: dXt dt = f(t;Xt;Wt) (3.1) Yt = g(t;Xt;Vt) (3.2) where Xt 2Rnx denotes the state vector, Yt 2Rny denotes the measurement vector, Wt 2 Rnw is a random vector representing the uncertainties in the model, and Vt 2Rnv denotes the random measurement error. In many applications, the noise from measurement can be assumed to be additive, and the problem can be formulated in a discrete manner as Xt+1 = ft(Xt;Wt) (3.3) Yt = gt(Xt) +Vt; (3.4) where fWtgt2N2Rnw and fVtgt2Nnf0g2Rnv are mutually independent white noises and the subscript t indexes the discrete time level at which the functions are evaluated. In data assimilation, the observation Yt arrives sequentially in time and the goal is to estimate the state vector Xt given the information of fYs; 0 0, there exists N0, such that j n(z) (z)j<" for all z2B and n>N0. 19 n converges to 2P(B) weakly and write limn!1 n ?= if limn!1h n;?i=h ;?i; 8?2Cb(B); where Cb(B) is the set of all continuous bounded functions on B. We will prove the weak convergence of (xtjIt) top(xtjIt), i.e., the convergence ofh (xtjIt);?i to hp(xtjIt);?i. To guarantee that the Bayes? formula in (3.7) is well de ned and can be ful lled in our algorithm, we make the following standing assumptions: (A1) For given It, the denominator in (3.7) (normalization constants) satis es Z B p(ytjxt)p(xtjIt 1)dxt > > 0 (A2) The conditional kernel densitiesp(xtjxt 1) andp(ytjxt) are uniformly continuous, bound- ed and strictly positive, i.e., given It; 0 0 such that @@xf 1t < , then sM at( M;N)2PC(B) and sM at( M)2 PU(B). Proof. For any x2B, by the de nition of sM and at we have sM at( M;N)(x) = 1M MX j=1 M;N x(x;j)t 1 and sM at( M)(x) = 1M MX j=1 M x(x;j)t 1 : 24 Since limN!1 M;N= M for each M2N, given any "> 0, there exists N0, such that for all z2B, j M;N(z) M(z)j<" for each M2N. Therefore, for all x2B, sM a t( M;N)(x) sM at( M)(x) = 1 M MX j=1 M;N x(x;j)t 1 M x(x;j)t 1 1M MX j=1 M;N x(x;j)t 1 M x(x;j)t 1 < 1M MX j=1 " = ": This proves that limN!1sM at( M;N)=sM at( M) for all M2N. We next prove that sM at( M;N) 2PC(B). In fact, for any " > 0 and z0 2B, since M;N 2PC(B), there exists > 0 such that when jz z0j< , j M;N(z) M;N(z0)j<": Fix arbitrary x0 2B, for any x2Bsatisfyingjx x0j< = , using that ft(x(x;j)t 1 ;w(j)t 1) = x and ft(x(x0;j)t 1 ;w(j)t 1) = x0 we have x(x;j)t 1 x(x0;j)t 1 =jf 1t (x;w(j)t 1) f 1t (x0;w(j)t 1)j @ @xf 1 t jx x0j< ; and thus sM a t( M;N)(x) sM at( M;N)(x0) 1 M MX j=1 M;N x(x;j)t 1 M;N x(x0;j)t 1 <": It remains to show that sM at( M)2PU(B). In fact, given any x1;x2 2B, sM at( M)(x1) sM at( M)(x2) = 1M MX j=1 M x(x1;j)t 1 M x(x2;j)t 1 : 25 For any " > 0, from the uniformly continuity of M, there exists > 0, such that for any z1;z2 2B with jz1 z2j< , j M(z1) M(z2)j<". Let ~ = , then jx1 x2j< ~ implies that x(x1;j)t 1 x(x2;j)t 1 = f 1t (x1;w(j)t 1) f 1t (x2;w(j)t 1) @ @x(f 1 t ) jx1 x2j< : Hence sM a t( M)(x1) sM at( M)(x2) = 1M MX j=1 M x(x1;j)t 1 M x(x2;j)t 1 <": The proof is complete. Lemma 3 For f M;Ng1M;N=1 2PC(B) and M 2PU(B) with limN!1 M;N= M for each M2N, it holds that lim N!1 TN( M;N)= M; 8M2M: (3.24) Proof. For any xt2B, TN( M;N)(xt) M(xt) TN( M;N)(xt) TN( M)(xt) + TN( M)(xt) M(xt) : (3.25) Since limN!1 M;N= M, for any " > 0, there exists N1 = N1(M) > 0 such that when N >N1, j M;N(xt) M(xt)j< "2: Thus because of the linearity of TN we have TN( M;N)(xt) TN( M)(xt) = TN( M;N M)(xt) < " 2: (3.26) For the second term on the right hand side of inequality (3.25), since TN is the linear interpolation operator and M is uniformly continuous, for any " > 0, there exists N2 = 26 N2(M) > 0 such that when N >N2 we have TN( M)(xt) M(xt) < " 2: (3.27) In summary letting N0 = maxfN1;N2g we have by (3.25), (3.26) and (3.27) that for any "> 0, TN( M;N)(xt) M(xt) <"; 8N >N 0; 8xt2B; 8M2N: (3.28) The proof is complete. We next prove the weak convergence of the operator M;Nt to t. Letting M;Nt and t be the composition operators de ned as above, we have the following theorem. Theorem 3.1 (Local convergence) Assume that the transition kernel p(xtjxt 1) is Feller and p(ytjxt) is bounded, uniformly continuous, and strictly positive. Also assume that @@xf 1t is bounded. Then, for any f M;Ng1M;N=1 2PC(B) and M; 2PU(B) with limN!1 M;N= M for each M2N and limM!1 M ?= , it holds that lim M!1 lim N!1 M;Nt ( M;N) ?= t( ): (3.29) Proof. Given limM!1 M ?= , by Lemma 1, lim M!1 sM at( M) ?= at( ): (3.30) Given limN!1 M;N= M for eachM2N, by Lemma 2 we have limN!1sM at( M;N)=sM at( M) and sM at( M;N) 2 PC(B), sM at( M) 2 PU(B). Thus by letting M;N := sM at( M;N) and M := sM at( M) in Lemma 3 we get lim N!1 TN sM at( M;N)=sM at( M): (3.31) 27 Equations (3.30) and (3.31) together give lim M!1 lim N!1 TN sM at( M;N) ?= at( ): Therefore it follows directly from (3.19) that lim M!1 lim N!1 bt TN sM at( M;N) ?= bt at( ): The proof is complete. To prove the global weak convergence result, we also need the following Lemma. Lemma 4 Assume p(ytjxt) is bounded, uniformly continuous, and strictly positive. For f M;Ng1M;N=1 2PC(B) and M 2PU(B) with limN!1 M;N= M for each M 2N, if there exists a 0 > 0 such that RBp(ytjxt) M(xt)dxt 0, then we have lim N!1 bt( M;N)=bt( M)2PU(B); 8M2N: Proof. Since limN!1 M;N= M for each M 2N, then for any 0 < " < 02 , there exists N0, such that when N >N0, j M;N(xt) M(xt)j<"; 8xt2B and 8M2N: It then follows that RBp(ytjxt)( M;N(xt) M(xt))dxt <" and Z B p(ytjxt) M;N(xt)dxt > Z B p(ytjxt) M(xt)dxt "> 02 : 28 Thus for any xt2B when N >N0 we have jbt( M;N)(xt) bt( M)(xt)j = p(ytjxt) M;N(xt) R Bp(ytjxt) M(xt)dxt M(xt) R Bp(ytjxt) M;N(xt)dxtR Bp(ytjxt) M;N(xt)dxt R Bp(ytjxt) M(xt)dxt R Bp(ytjxt) ( M(xt) M;N(xt))dxtR Bp(ytjxt) M;N(xt)dxt R Bp(ytjxt) M(xt)dxt + M(xt) M;N(xt)R Bp(ytjxt) M(xt)dxt < 4 20 + 2 0 ": (3.32) Therefore limN!1bt( M;N)=bt( M). It remains to show that bt( M)2PU(B). In fact, by the de nition of bt, we have for any x(1)t ;x(2)t 2B, bt( M)(x(1)t ) bt( M)(x(2)t ) 2 0 p(ytjx(1)t ) M(x(1)t ) p(ytjx(2)t ) M(x(2)t ) : From the uniformly continuity property and the boundedness of M and p(xtjxt 1) that for any " > 0, there exists > 0, such that when x(1)t ;x(2)t 2B with jx(1)t x(2)t j< , we have j M(x(1)t ) M(x(2)t )j< "2 and jp(x(1)t ) p(x(2)t )j< "2. Thus, bt( M)(x(1)t ) bt( M)(x(2)t ) 0 such that RBp(ytjxt)sM p(xtjIt 1)dxt > 0 for M su cient large. Applying Theorem 3.1 to the context of ltering problems, we can obtain the weak convergence of our implicit ltering simulation to the Bayesian optimal lter. Our main result of this work is stated in the following theorem. For simplicity, we de ne two new 29 operators Mt and M1:t to be Mt = bt sM at and M1:t = t t 1 1: Theorem 3.2 (Global convergence) Assume that the transition kernel p(xtjxt 1) is Feller and p(xtjxt 1) is bounded, uniformly continuous, and strictly positive. Also assume that @ @xf 1 t is bounded. Then lim N!1 M;N1:t (p0j0) = M1:t(p0j0) 8M2N; and lim M!1 M1:t(p0j0) ?= 1:t(p0j0); which implies that lim M!1 lim N!1 tjt ?= ptjt: Proof. To prove Theorem 3.2, we use induction method. (1) t = 1: choose M;N = M = = p0j0 in equation (3.29). It is obviously that lim N!1 M;N= M; lim M!1 M ?= ; and M = p0j0 is uniformly continuous. By Lemma 2, Lemma 3 and Lemma 4 and Note 1, lim N!1 M;N1 (p0j0) = M1 (p0j0)2PU(B); 8M2N: By Lemma 1 and the continuity of bt, we have lim M!1 M1 (p0j0) ?= 1(p0j0) = p1j1: It then follows from Theorem 3.1 that lim M!1 lim N!1 M;N1 (p0j0) ?= 1(p0j0) = p1j1: 30 (2) Assume that lim N!1 M;N1:t 1(p0j0) = M1:t 1(p0j0)2PU(B) 8M2N; and lim M!1 M1:t 1(p0j0) ?= 1:t 1(p0j0): We choose M;N = M;N1:t 1(p0j0) and = 1:t 1(p0j0) = pt 1jt 1 in equation (3.29) and M = M1:t 1(p0j0) in Theorem 3.1. From the assumption, lim N!1 M;N= M and lim M!1 M ?= : By Lemma 2, Lemma 3, Lemma 4 and Note 1 that lim N!1 M;N1:t (p0j0) = lim N!1 bt TN sM at( M;N1:t 1(p0j0))= M1:t(p0j0)2PU(B): (3.33) By Lemma 1 and the continuity of bt we have lim M!1 M1:t(p0j0) ?= ptjt: (3.34) Therefore it follows from Theorem 3.1 that lim M!1 lim N!1 M;Nt ( M;N1:t 1(p0j0)) ?= t(pt 1jt 1) = 1:t(p0j0) = ptjt: The proof is complete. 3.3 Numerical experiments In this section, we present two numerical examples to demonstrate the e ciency of our method. The rst example involves a one dimensional nonlinear system and measurement equation while the second is a 2-D bearing-only tracking problem. We shall compare our method with the standard EKF and particle lter. Here the particle lter we are using is 31 the sequential important sampling with resampling (SIR). Example 1 Consider the following nonlinear model xk = 40 tan(xk 1 + 10) + 50wk 1; yk = 40 xk2000 +x k +vk; (3.35) where wk and vk are two independent zero-mean white noise processes with variance 1:0, yk is the noise perturbed observation of xk. The initial position is taken to be x0 = 2 and Figure 3.1 shows a 50 step realization of the state equation in model (3.35). 0 5 10 15 20 25 30 35 40 45 50?600 ?400 ?200 0 200 400 600 800 1000 k Xk Figure 3.1: Original Position Figure 3.2, Figure 3.3 and Figure 3.4 are the simulation results obtained by using EKF and particle lter and our implicit particle lter method, respectively. The true state is represented by blue diamonds while simulation results are given as red \stars" and connected by solid lines. The prior pdf p(x0) is initialized with the standard normal distribution with the mean value x0 and the variance 1:0. 32 In particle lter method, we use 500 particles (sample points) to represent the pdf and in our implicit lter, we use 100 nodes to partition the region and the number of Monte-Carlo samples is M = 10. 0 5 10 15 20 25 30 35 40 45 50?1500 ?1000 ?500 0 500 1000 1500 Real State EKF Figure 3.2: Extended Kalman Filter 0 5 10 15 20 25 30 35 40 45 50?600 ?400 ?200 0 200 400 600 800 1000 Real State Particle Filter Figure 3.3: Particle Filter Form the three gures, one can see that when the variation between two consecutive points is not very large, all three methods produce very accurate approximations to the true state. On the other hand, when the true state has very large variations at some time steps, i.e., the state variable has a large jump from its previous state, both EKF and particle lter 33 fail to produce accurate approximations. However, our implicit particle algorithm still pro- duces accurate estimations at these points. 0 5 10 15 20 25 30 35 40 45 50?800 ?600 ?400 ?200 0 200 400 600 800 1000 Real State Implicit Algorithm Figure 3.4: implicit Algorithm Example 2 In this example, we consider the following bearing-only tracking problem. dX1(t) = X2(t)dt+ X1(t)(X 1(t))2 + (X2(t))2 + 1dW1(t); dX2(t) = X1(t)dt+ X2(t)(X 1(t))2 + (X2(t))2 + 2dW2(t) (3.36) where W1(t) and W2(t) are two independent Brownian Motions. This stochastic dynamical system may serve to model the motion of a ship which moves with a constant radial and angular velocity, perturbed by a white noise. The observations are collected by a detector located at a platform with time intervals of length = 0:05 and the data are angular measurements corrupted by noise. 34 To approximate the state variables X = (X1;X2), we discretize the dynamical system (3.36) in time and obtain a discrete nonlinear ltering problem. Let xk = (x1k;x2k). We have the discrete system model x1k = x1k 1 x2k 1 + x1k 1(x1 k 1)2+(x 2 k 1)2 + 1p w1k 1 x2k = x2k 1 + x1k 1 + x2k 1(x1 k 1)2+(x 2 k 1)2 + 2p w2k 1: (3.37) The mathematical formula for the measurement equation is given by yk = arctan(x 2 k x 2 platform x1k x1platform) + p v k; (3.38) where xplatform = (x1platform;x2platform) is the location of the platform where a detector is placed. ?80 ?60 ?40 ?20 0 20 40?80 ?60 ?40 ?20 0 20 40 60 X Y platform Figure 3.5: Target Positions In the numerical simulations the model parameters are chosen as = 5, = 2 and 1 = 2 = 8. Figure 3.5 gives the target path in the x-y plan, with the position of the target at each time step shown by a diamond. The location of the detector platform is chosen as xplatform = ( 15; 15), marked by a red box. 35 The problem is initialized with a best guess of the target position at the initial time, which is (x10;x20) = (0:5;0:5). In this example, we use 8000 particles (sample points) to represent the pdf in the particle lter method and in the implicit lter, we use 1600 nodes to partition the region and the number of Monte-Carlo samples is M = 10. p Figure 3.6 shows the simulation results of observation angle using EKF, particle lter and our implicit particle lter method. From this gure, one can see that both particle lter and the implicit particle lter produce good approximation for the relative observation angle of the target. Although the EKF provides the trend of the movement of the target, the estimation is a few steps delayed from the true target observation angle. 0 10 20 30 40 50 60 70 80?100 ?80 ?60 ?40 ?20 0 20 40 60 80 100 RealImplicit Algorithm Particle FilterEKF Figure 3.6: Comparison result for the observation angle Figure 3.7 shows the results of system state simulations using EKF while Figure 3.8 compares the performance between particle lter and the implicit particle lter. Clearly both the particle lter and implicit particle lter outperform EKF. While the accuracy of the particle lter and the implicit particle lter are very close to each other at the initial stage, the implicit particle lter becomes more accurate at the nal stage of time interval. 36 ?80 ?60 ?40 ?20 0 20 40?80 ?60 ?40 ?20 0 20 40 60 Real StateEKF Figure 3.7: Simulation result of EKF ?80 ?60 ?40 ?20 0 20 40 60?100 ?80 ?60 ?40 ?20 0 20 40 60 Real Implicit Algorithm Particle Filter Figure 3.8: Comparison result between Particle Filter and implicit Algorithm 37 Chapter 4 A Hybrid Sparse Grid Approach for Nonlinear Filtering Problems Based on Adaptive-Domain of the Zakai Equation Approximations In this Chapter, we develop an e cient hybrid sparse grid approach for nonlinear l- tering problems based on numerical approximations of the Zakai equation. Consider the following stochastic di erential system that combines an equation for the \state" and for the \observation" de ned on the probability space ( ;F;P) 8> < >: dXt = b(Xt)dt+ (Xt)dWt; (state) dYt = h(Xt)dt+dBt: (observation) (4.1) Here fXt 2Rd;t 0g and fYt 2Rr;t 0g are two stochastic processes, fWt;t 0g and fBt;t 0g are independent Brownian Motions in Rp and Rr, with covariance matrices CW (identity) and CB, respectively, and the given initial value X0 is independent of Wt and Bt with probability distribution u0(x)dx. 4.1 Nonlinear Filtering Problems and Zakai Equations Now, we outline the derivation of the Zakai equation and its relationship to the nonlinear ltering problem (4.1). Throughout this Chapter, we assume that the coe cients b : Rd! Rd, : Rd !Rd p and h : Rd !Rr in (4.1) are globally Lipschitz continuous functions. Denote (t) := exp Z t 0 h (Xs)dYs 12 Z t 0 jh(Xs)j2ds ; 38 then the measure ~P de ned by ~P = (t)dP is also a probability measure on ( ;F) equivalent to P . Furthermore, in the probability space ( ;F; ~P), Yt is a Browanian motion independent of Xt ( for details, see [82] ). Assuming that u = u(t;x) is the conditional density function of the state Xt given an observed path Yt, then the optimal ltering solution is given by (see [82, 85]) E( (Xt)jYt) = R (x)u(t;x)dx R u(t;x)dx : (4.2) Under regularity assumptions for the coe cients b and h given above, u satis es the following stochastic partial di erential equation, known as Zakai equation du(t;x) = L u(t;x)dt+h (x)u(t;x)dYt; x2Rd; (4.3) with the initial value u(0;x), and L the in nitesimal generator associated with the state process Xt such that Lu = 12 dX i;j ( )i;j @ 2u @xi@xj + dX i=1 bi @u@x i ; (4.4) and \ " is the transpose operator which transforms Lu to be L u = 12 dX i;j @2( )i;ju @xi@xj dX i=1 @biu @xi : (4.5) The goal of the Zakai lter method is to obtain numerical solutions of the Zakai equation (4.3). However, there are several challenges in the construction of an e cient numerical algorithm for the Zakai equation: (i) high-dimensionality of the state equations; (ii) low regularity of the solution; and (iii) unbounded solution domain. In the next two sections we construct a hybrid algorithm combining the ideas of split-up nite di erence method, sparse grid interpolation and the importance sample approach to overcome these obstacles. 39 4.2 Hierarchical Local Sparse Grid Interpolation In this section, we introduce a sparse grid interpolation constructed from a local hier- archical basis which will be used in the nite di erence approximation of the Zakai equation in the spatial domain. 4.2.1 Standard hierarchical sparse grid ipnterpolation Assume that we have the following one dimensional interpolation formula at our disposal: Qi(u) = miX j=1 u(xij) ij(x); x2R; (4.6) where i2N is the resolution level of the interpolant Qi, mi is the number of grid points on level i, xij and ij(x) for j = 1;:::;mi are the interpolation points and the corresponding basis functions, respectively. In the context of linear hierarchical interpolation, mi, xij and ij in the standard interval [ 1;1] for i2N;j = 1;:::;mi are de ned by mi = 8 >< >: 1; if i = 1; 2i 1 + 1; if i> 1; (4.7) xij = 8 >>< >>: 0; for j = 1; if mi = 1; 2(j 1) mi 1 1; for j = 1;:::;mi; if mi > 1; (4.8) and for i = 1, 11 = 1; for i> 1 and j = 1;:::;mi, ij = 8 >>< >>: 1 mi 12 jx xijj; if jx xijj< 2m i 1 ; 0; otherwise. (4.9) Note that the nodal basis function ij has local support [xij 21 i;xij + 21 i]. 40 In the multi-dimensional case, i.e. d> 1, the tensor-product interpolatant is Qi 1 Qid (u) = mi1X j1=1 midX jd=1 u xi1j1; ;xidjd ij(x); (4.10) where ij = Qdk=1 ikjk. Clearly, the above product requires di=1mi function values, which is computationally prohibitive when d is large. The sparse gird interpolation [17] is a linear combination of a series of tensor-product interpolants, each of which is de ned on a coarse grid with di erent resolutions in di erent dimensions, i.e., IL;d(u) = X L d+16jij6L ( 1)L jij d 1 q jij Qi1 Qid (u); (4.11) where L d, the multi-index i = (i1;:::;id) and jij = i1 + + id. Here, ik(k = 1;:::;d) is the level of the tensor-product interpolant Qi1 Qid along the kth direction. The Smolyak algorithm builds the interpolant by adding a combination of all tensor-product interpolants satisfying L d + 1 jij L. The structure of the algorithm becomes clearer when one considers the incremental interpolant, i given in [17] Q0(u) = 0; i =Qi(u) Qi 1(u): (4.12) The sparse grid interpolant (4.11) is then equivalent to IL;d(u) = X jij6L ( i1 id) =IL 1;d(u) + X jij=L ( i1 id)(u): (4.13) The corresponding sparse grid associated with IL;d(u) is represented by HL;d = [ L d+16jij6L ( i1 id); (4.14) 41 where i denotes the set of interpolation points used by Qi. According to (4.13), to extend the Smolyak interpolantIL;d(u) from level L 1 to L, one only needs to evaluate the function at the incremental grid HL;d de ned by HL;d = [ jij=L ( i1 id); (4.15) where ij = ij n ij 1, j = 1;:::;d. According to the nested structure of the one- dimensional hierarchical grid de ned by (4.8), it is easy to see that i 1 i and i = in i 1 has mi = mi mi 1 points. By consecutively numbering the points in i, and denoting the jth point of i as xij, the incremental interpolant in (4.12) can be represented by (see [17, 55] for details) i(u) = mi X j=1 ij u(xij) Qi 1(u)(xij) ; (4.16) where !ij = u(xij) Qi 1(u)(xij) is de ned as the one-dimensional hierarchical surplus on level i. This is just the di erence between the values of the interpolating polynomials and the function evaluated at xij. From (4.16), the hierarchical sparse grid interpolant (4.13) can be rewritten as IL;d(u) =IL 1;d(u) + X jij=L ( i1 id)(u) =AL 1;d(u) + X jij=L j2Bi !ij ji(x) = X jij6L X j2Bi !ij ji(x); (4.17) where the multi-index set Bi is Bi = j2Nd : xikjk 2 ik for jk = 1;:::;mik ;k = 1;:::;d ; (4.18) 42 and the surpluses !ij are !ij = u(xi1j1;:::;xidjd) IL 1;d(u)(xi1j1;:::;xidjd): (4.19) As proved in [17], for smooth functions, the hierarchical surpluses tend to zero as the in- terpolation level tends to in nity. On the other hand, the magnitude of the surplus is a good indicator about the smoothness of the interpolated function. In general, the larger the magnitude, the stronger the underlying discontinuity. For a bounded box domain D Rd : D= [a;b] := di=1[ai;bi]; (4.20) we rst transform it to [ 1;1]d through a simple linear transform. The corresponding sparse grid interpolation is then de ned according to (4.11). 4.2.2 Hierarchical sparse grid approximation of bell-shaped cpurves In many nonlinear ltering problems in practical applications, the conditional target state PDF resembles a \bell-shaped" Gaussian curve or surface, if not exactly Gaussian. In such cases, standard hierarchical sparse grid interpolations may lead to large errors. This is especially the case for two or higher dimensional problems. As a demonstration, we consider a Gaussian-type function U(x) := dY i=1 exp ( 12 x i 2) ; x2Rd; (4.21) where 2R and 2R+. From [17], we know that the L2 error between the function U and the standard sparse grid approximation U is kU UkL2 2 jUj2;212d 2 2L A(d;L); (4.22) 43 where L2N+ represents the level of hierarchical sparse grid with A(d;L) = L d (d 1)! +O(k d 2); and jUj2;2 =kD2UkL2: It can be shown that jUj2;2 in (4.22) is bounded by (see [64] for detailed derivation) jUj2;2 1 4d(maxfxi g2d)kUkL2: (4.23) We can see from the above estimate that when the constant is small, the right hand side 100 101 102 103 104 10510 ?3 10?2 10?1 100 101 # hierarchical sparse?grid points Interpolation error 1?Dim 2?Dim 3?Dim 4?Dim 5?Dim Figure 4.1: The interpolation error for regular sparse grid approximation up to 5 dimensions. becomes exponentially large as d increases. Here, we consider a special case of (4.21) with = 0 and = 0:15, i.e., U(x) = dY i=1 exp x 2 i 0:045 : (4.24) Figure 4.1 shows the interpolation errors using the L2-norm for regular sparse grid approxi- mations of U. We can see that as the dimension d increases, the interpolation errors barely decrease even, though the number of interpolation points has increased signi cantly. 44 100 101 102 103 104 10510 ?4 10?3 10?2 10?1 100 101 # hierarchical sparse?grid points Interpolation error 1?Dim 2?Dim 3?Dim 4?Dim 5?Dim Figure 4.2: The interpolation error for logarithmic sparse grid approximation up to 5 dimen- sions. One way of alleviating this poor performance of the sparse grid interpolant when ap- proximating bell-shaped-functions is to utilize logarithmic interpolation (see [64]), in which we take the logarithm of U, i.e. V := log(U), and build the approximation V using the standard sparse grid approach. Then we obtain the approximation of U by U = e V . Figure 4.2 shows the absolute interpolation errors measured in L2-norm for the logarithmic sparse grid approximation of function U de ned in (4.24). Compared with Fig. 4.1, we can see from Figure 4.2 that the convergence is improved as the dimension increases. A visual demonstration of the e ciency of the logarithmic sparse grid interpolation is shown in Fig. 4.3, where we plot the level 6 approximation for a marginal distribution of both U and log(U) in d = 4 dimensions. 4.3 Hybrid Approach for Numerical Solution of the Nonlinear Filtering Prob- lem. In this section we describe our hybrid numerical algorithm for the solution of the Zakai equation (4.3). 45 ?1 0 1 ?1 0 10 2 4 6 8 x1 (a) y1 ?1 0 1 ?1 0 1?5 0 5 10 15 x1 (b) y1 ?1 0 1 ?1 0 1?30 ?20 ?10 0 10 x1 (c) y1 ?1 0 1 ?1 0 10 2 4 6 8 x1 (d) y1 Regularsparsegrid approximation V=Log(U);Regularsparsegrid approximationforV U=Exp(V) Figure 4.3: (a) is the x1x2-marginal surface of U with x3 = x4 = 0 while (b) is its interpolate approximation obtained by applying the regular sparse grid approximation to function U. We can see from the Figure that the approximation is quite poor and signi cant oscillations occur at the bottom of the surface. On the other hand, from (d) one can see that the logarithmic interpolation approximation described above is far more accurate. 4.3.1 Adaptive selection of solution domains Since the Zakai equation (4.3) is de ned on the whole space Rd, it is essential to choose a proper bounded domain to achieve an accurate numerical approximation. Motivated by importance sampling and particle lter methods, we adaptively select a hypercube at each time step according to an estimation of the density function of the solution at the next step. In particular, let Rt be a partition of [0;T] such that: Rt =ftnjtn2[0;T];tn < >: i ^un(x+eihi) un(x)h i if i 0 iun(x) ^un(x eihi)h i if i < 0 ; where ei is the unit vector in the ith coordinate direction and hi is a properly chosen meshsize in the ith coordinate direction. To approximate second order partial derivatives at the sparse grid point x2HL;dn+1 with given coe cient function 2Rd d, we use central di erences to obtain i;i @ 2un @xi@xi(x) i;i ~D2x ixiun(x) := i;j ^un(x+eihi) 2un(x) + ^un(x eihi) h2i and i;j @ 2un @xi@xj(x) i;j ~D2x ixjun(x) := 8 >>> >>>> >>> >>< >>> >>> >>> >>>: i;j 2hi ^u n(x+eihi +ejhj) ^un(x+eihi) hj ^un(x+ejhj) un(x) hj +un(x) ^un(x ejhj)h j ^un(x eihi) ^un(x eii ejhj)h j ;if i;j 0; i;j 2hi ^u n(x+eihi) ^un(x+eihi ejhj) hj un(x) ^un(x ejhj) hj + ^un(x+ejhj) un(x)h j ^un(x eihi +ejhj) ^un(x eihi)h j ;if i;j < 0: 49 With the above nite di erence operators in hand, we de ne the nite di erence approxi- mation of the Fokker-Planck equation (4.28) on sparse grid HL;dn+1 as follows. un+ 1 2 (x) = un(x) +L nun(x) tn; x2HL;dn+1; (4.29) where L nun(x) = 12 dX i;j @2( ) i;j @xi@xj un(x) + @( )i;j @xi ~Dx jun(x) + @( )i;j @xj ~Dx iun(x) +( )i;j ~D2xixjun(x) o dX i=1 @bi @xiun(x) +b i ~Dx iun(x) : Update Step In the update step, we use the new observation Zn and the Bayes formula [?] to update the prior un+ 1 2 to the posterior un+1 as follows. un+1(x) = Cn n(x;Zn)un+ 1 2 ; x2HL;dn+1; (4.30) where Cn is a normalization factor and function n is de ned by n(x;Zn) = exp tn2 jZn h(x)j2R : The normj jR is de ned byj j2R = R 1 , where R is the covariance matrix offBt;t 0g in (4.1) ( Please see Page 2, [85] for more details). Finally following the procedure described in Section 3, we derive the logarithmic sparse grid interpolation un+1 = un+1(x); x2 Rd using its values on the sparse grid HL;dn+1. We summarize our hybrid sparse grid adaptive-domain splitting-up nite di erence al- gorithm as follows: 50 Step 1: Input u0 as the initial value of the solution u of the Zakai equation (4.3). Step 2: For n = 0 ;NT 1, 1 Compute dynamic domainDn+1 for the solution un+1 using the importance sampling method. 2 Generate sparse grid HL;dn+1 on the solution domain Dn+1. 3 Evaluate un+1 on the sparse grid HL;dn+1 by using nite di erence scheme (4.29). 4 Extend the solution un+1 to the whole space Rd through the logarithmic sparse grid interpolation described in Section 3. Step 3: Normalization. Remark 2 Since we use an explicit nite di erence scheme to solve equation (4.28), time step tn must satisfy the following stability condition max 0 n NT 1 tn 1dX i=1 j( T)i;ij+jbijhi h2i : 4.4 Numerical Experiments In this section, we present three numerical experiments to demonstrate the e ectiveness of our new numerical algorithm, for solving nonlinear ltering problems. Example 1 In the rst example, we use a two dimensional nonlinear ltering problem to illustrate the accuracy of the selection process of the dynamic solution domain Dn. To see this, we consider the following dynamical system dXt = (40;2 (10t)2)T dt+ 0:5 dWt; (4.31) 51 15 20 25 30 35 40 45 50 55 6015 20 25 30 35 40 45 50 55 60 X Y Figure 4.4: Target Trajectory and Adaptive Solution Domain. The red curve shows the real target state. The blue points are actual states of the target and the blue boxes are the corresponding solution domains. where Wt is a two-dimensional Brownian Motion and the initial state is given by X0 = (30;30)T. The observation process is given by dYt = q (X1t 20)2 + (X2t )2 dt+dBt; (4.32) which measures the perturbed distance between the target state and a reference point P = (20;0), and Bt is a one-dimensional Brownian motion independent of Wt. In this numerical simulation, we take T = 0:4 and use an uniform partition in time with stepsize tn = 0:005. The initial value is given by u0 N(X0; ); a normal distribution with mean X0 and standard deviation = (1;0:5)T. In the adaptive solution domain selection process, we choose the sample size M = 500 and the parameter in (4.25) as = 4. Figure 4.4 shows the trajectory of the target state and solution domain Dn for n = 1;20;50;70;80. In Figure 4.5 we show solution domain Dn with the corresponding contour plot of target 52 15 25 35 45 5520 25 30 35 40 45 1e?008 1e?008 1e?005 1e?005 0.005 X Y Real StatePDF (a) 15 25 35 45 5520 25 30 35 40 45 5e?005 5e?0050.001 0.0010.01 0.010.050.1 X Y Real StatePDF (b) 15 25 35 45 5520 25 30 35 40 45 5e?005 5e?005 0.0010.0010.010.01 0.050.1 X Y Real StatePDF (c) Figure 4.5: Target state PDF at time step: (a) n=1; (b) n=50; (c)=80; in the corresponding solution domain state PDF for n = 1;50;80. From this gure one can see the that solution domains are extremely accurate in approximating the high density area of un. 53 Example 2 In this example, we consider the following nonlinear ltering problem: 8> < >: dXt = (10;6 (sinX1t + 2))T dt+dWt; (state) dYt = (X1t;X2t )T dt+dBt: (observation) (4.33) In this numerical experiment, we let T = 0:5 and use a uniform partition in time with stepsize tn = 0:01. An example of the signal trajectory is shown in gure 4.6a. The initial value is given by u0 N(X0; ); a normal distribution with mean X0 = (2;0)T and standard deviation = (0:5;0:5)T. We also choose the hierarchical sparse grid level as L = 6. In the adaptive solution domain selection process, the sample size and parameter are given by M = 500 and = 4, respectively. Figure 4.6b shows the comparison of the estimated values of the target state between the particle lter and sparse grid Zakai lter. In the particle lter simulation, we use 8000 particles to represent the PDF of the target state. The black dashed line shows the trajectory of the real target state, the red triangles and blue dots show the estimate target state (the mean of the estimate posterior PDF) obtained by using the particle lter and the Zakai lter respectively. As we can see from Figure 4.6b, our hybrid approach is more accurate than the PFM for a longer period of time. To further examine the performance of the hybrid sparse grid Zakai lter method, in Figure 4.7a and Figure 4.7b we plot the marginal probability density functions at time T = 0:5, obtained by using the particle lter and our hybrid with respect to xy-coordinates. From the plots we observe that the convergence of the particle lter as the particle size increases. Moreover, the PDE obtained by our hybrid sparse grid Zakai lter is very close to the one obtained by the particle lter with 160;000 particles. However, in table 4.1 we can see that with 160;000 particles the particle lter is far more costly in terms of CPU time than our method. Finally, we also show the con dence bands in Figure 4.8. The blue dashed 54 2 3 4 5 6 7 80 1 2 3 4 5 6 X Y Target Trajectory Starting Point (a) Signal trajectory: Example 2 2 3 4 5 6 7 80 1 2 3 4 5 6 X Y Target Trajectory Particle Filter Zakai Filter (Sparse Grid) (b) Comparison of estimated mean of the target state between particle lter and sparse grid Zakai lter Figure 4.6: 2D nonlinear ltering problem curves show the real target trajectory with respect to xy-coordinates, respectively, the red 55 2 3 4 5 6 7 8 9 10 11 120 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 x PF: 2,000 PF: 8,000PF: 40,000 PF: 160,000Zakai Filter (a) Marginal distribution with respect to X-coordinate 1 2 3 4 5 6 7 8 9 10 110 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 y PF: 2,000 PF: 8,000PF: 40,000 PF: 160,000Zakai Filter (b) Marginal distribution with respect to Y-coordinate Figure 4.7: Marginal distributions curves represent the estimate of posterior means, and the green dashed curves represent the estimated 95% con dence bands. 56 0 10 20 30 40 500 1 2 3 4 5 6 7 8 9 10 n X n Real State ZF(Sparse Grid) Estimated StateZF(Sparse Grid) 95% probability region (a) 0 10 20 30 40 50?1 0 1 2 3 4 5 6 7 8 9 n Y n Real State ZF(Sparse Grid) Estimated StateZF(Sparse Grid) 95% probability region (b) Figure 4.8: Sparse grid Zakai lter estimate of posterior mean with 95% probability region . (a) Estimate of posterior mean and 95% probability region: X-coordinate (b) Estimate of posterior mean and 95% probability region: Y-coordinate Table 4.1: Comparison of computing costs Particle lter CPU time (seconds) 2,000 particles 1:49 8,000 particles 32:67 40,000 particles 516:91 160,000 particles 7622:29 Sparse grid Zakai lter 20:45 Example 3 In this example, we consider the following \ bearing-only " tracking problem given by: dXt = bdt+ dWt; (4.34) where Xt = (x;y;u;v)Tt is a four-dimensional vector which models the movement of a target ship sailing on the sea plane (x y plane). Here (x;y) and (u;v) are the position and velocity 57 components respectively, the vector b = (u;v;0;0)Tt , the covariance matrix 2 is de ned by 2 = 0 BB BB BB B@ 21 0 0 0 0 22 0 0 0 0 23 0 0 0 0 24 1 CC CC CC CA ; (4.35) and Wt is a four-dimensional Brownian motion. To estimate the target state, a passive sonar 0 5 10 15 20 25 300 5 10 15 20 25 30 x y Taget Ownship Observation angle ? Figure 4.9: Bearing-only tracking is located on an observation ship, denoted \ownship". The observation process is given by: dYt = h(Xt)dt+dBt; 58 where the observation function h is the angle h(Xt) = arctan y t yobst xt xobst ; andBt is a one-dimensional Brownian motion independent fromWt. Here, Xobst = xobs;yobs Tt describes the movement of the ownship given by dXobst = bobsdt. 5 10 15 20 256 8 10 12 14 16 18 x y Target Trajectory Particle Filter Zakai Filter (Sparse Grid) Figure 4.10: Comparison of estimated mean of the target state between particle lter and sparse grid Zakai lter In the numerical simulations, we choose 1 = 2 = 0:75, 3 = 4 = 0:05 in (4.35) and bobs = (8;0)T; thus the movement of the ownship is along the Y-axis with a constant speed. In addition, we set T = 1, the time partition tn = 0:005, initial target state X0 = (2;8;20;8)T, and the initial PDF of the target sate N( X; ), where X = (3;9;19:8;7:9)T and = (0:75;0:75;0:5;0:25)T. 59 20 40 60 80 100 120 140 160 180 2000 5 10 15 20 25 n x Target Trajectory ZF(SG) Estimated stateZF(SG) 95% probability region (a) 20 40 60 80 100 120 140 160 180 2006 8 10 12 14 16 18 20 n y Target Trajectory ZF(SG) Estimated stateZF(SG) 95% probability region (b) 20 40 60 80 100 120 140 160 180 20016 17 18 19 20 21 22 23 24 25 26 n u Target Trajectory ZF(SG) Estimated stateZF(SG) 95% probability region (c) 20 40 60 80 100 120 140 160 180 2006 7 8 9 10 11 12 n v Target Trajectory ZF(SG) Estimated stateZF(SG) 95% probability region (d) Figure 4.11: Sparse grid Zakai lter estimate of posterior mean with 95% probability region. (a) Estimate of posterior mean of the location x and 95% probability region. (b) Estimate of posterior mean of the location y and 95% probability region. (c) Estimate of posterior mean of the velocity u and 95% probability region. (d) Estimate of posterior mean of the velocity v and 95% probability region The target-observer plane (x y plane) is illustrated in Figure 4.9. The red dot shows the initial position of the target ship while the blue triangle shows the initial position of our ownship. The dashed red curve gives a possible trajectory of the target ship and the blue arrow describes the movement of our ownship. Figure 4.10 shows the comparison of the estimated mean values of the relative target position with respect to the ownship, in the target-observer plane. For the sparse grid Zakai 60 18 19 20 21 22 23 24 25 26 270 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 x PF: 20,000 PF: 40,000PF: 80,000 PF: 160,000Zakai Filter (a) Marginal distribution of the target state on x-coordinate 13 14 15 16 17 18 19 20 210 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 y PF: 20,000 PF: 40,000PF: 80,000 PF: 160,000Zakai Filter (b) Marginal distribution of the target state on y-coordinate Figure 4.12: Marginal distributions lter, we let the hierarchical level L = 6 and for the adaptive solution domain selection process, we set M = 1;000 and = 4. The estimate for the particle lter is obtained by using 160;000 particles. The actual trajectory of the target position relative to our ownship is given by the black curve. The red curve and blue curve show the estimate target state (the mean of the estimate posterior PDF ) obtained by using the particle lter and the Zakai 61 Table 4.2: Comparison of computing costs Particle lter CPU time (seconds) 20,000 particles 554:24 40,000 particles 1976:39 80,000 particles 7702:57 160,000 particles 32380:38 Sparse grid Zakai lter 587:14 lter respectively. We can see from Figure 4.10 that our hybrid sparse grid Zakai lter yields similar estimate results compared with the particle lter with 160;000 particles. In Figure 4.12a and 4.12b we compare the marginal probability distribution functions at time T = 1 with respect to xy coordinates using our hybrid sparse grid Zakai lter with the particle lter. Similar to Example 2, we see that the convergence of the particle lter with the particle size 160;000 to the approximate PDE is very close to the approximate PDE obtained by the hybrid sparse grid Zakai lter. In table 4.2, we show the comparison of computational from which one can see that the particle lter with 160;000 particles takes 60 times more computing time to achieve a similar PDF. We also plot the con dence curves in Figure 4.11 for further examination of the performance of sparse grid Zakai lter. The blue dashed curves show the real target trajectory with respect to xy coordinates. The red curves represent the estimate of posterior means, and the green dashed curves give the estimate 95% con dence curves ( 2 times of the estimated standard deviation). 62 Chapter 5 Numerical Algorithms for Backward Doubly Stochastic Di erential Equations and it?s Applications to Nonlinear Filtering Problems 5.1 FBDSDEs and SPDEs To derive the numerical algorithm and conduct its convergence analysis, we provide a brief introduction to forward backward doubly stochastic di erential equations (FBDSDEs) and the relationship between FBDSDEs and the SPDE. Let ( ;F;P) be a complete probability space and T > 0 be the terminal time,fWt;0 t Tg and fBt;0 t Tg be two mutually independent standard Brownian motions de ned on ( ;F;P) with their values in Rd and in Rl, respectively. Let N denote the class of P-null sets of F. For each t2[0;T], we de ne Ft :=FWt _FBt;T; where F s;t = f r s;s r tg_N the - eld generated by f r s;s r tg, and F t = F 0;t for a stochastic process . Note that the collection fFt;t 2 [0;T]g is neither increasing nor decreasing and it is not a ltration. For positive integer n 2 N, we de ne spaces M2(0;T;Rn) and S2([0;T];Rn) as follows. M2(0;T;Rn) := f?tj?t2Rn;E Z T 0 j?tj2dt<1; ?t2Ft; a.e. t2[0;T]g; and S2([0;T];Rn) :=f?tj?t2Rn;E( sup 0 t T j?tj2) <1; ?t2Ft; t2[0;T]g: 63 Let f : [0;T] Rk Rk d!Rk and g : [0;T] Rk Rk d!Rk l be jointly measurable such that for any (y;z)2Rk Rk d, ft(y;z)2M2(0;T;Rk); gt(y;z)2M2(0;T;Rk l): We assume moreover that there exist constants c > 0 and 0 < < 1 such that for any (!;t)2 [0;T], (y1;z1), (y2;z2)2Rk Rk l, jft(y1;z1) ft(y2;z2)j2 c(jy1 y2j2 +kz1 z2k2); kgt(y1;z1) gt(y2;z2)k2 cjy1 y2j2 + kz1 z2k2: From [61], under the above assumptions and standard conditions on b and , we know that there exists a pair of processesf(Yt;xs ;Zt;xs ); (t;x)2[0;T] Rdgwhich is the unique solution to the following FBDSDE: For (t;x)2R+ Rd Xt;xs = x+ Z s t b(Xt;xr )dr + Z s t (Xt;xr )dWr; t s T; (5.1) Yt;xs = h(Xt;xT ) + Z T s f(r;Xt;xr ;Yr;xr ;Zt;xr )dr + Z T s g(r;Xt;xr ;Yr;xr ;Zt;xr )d Br Z T s Zt;xr dWr; t s T; (5.2) where (Yt;xs ;Zt;xs ) 2 S2([0;T];Rk) M2(0;T;Rk l). Here d Br denotes the backward It^o integration, i.e., for a FBt;T adapted process Vt, and quasi-uniform time partitions : 0 = 64 t0 >>> >>< >>>> >>: ut(x) = h(x) + Z T t [12 dX i=1 @2us(x) @x2i +f(s;x;us(x);rus(x))]ds + Z T t g(s;x;us(x);rus(x))d Bs; x2 ; 0 t T; ut(x) = (t;x); on [0;T] @ : 5.2 Half Order Numerical Algorithms We rst study numerical algorithms for BDSDEs and assume that b 0 and 1 in (5.1). Thus we have X0;xs = x+Ws; x2 ;s2[0;T]; 66 and the elliptic partial di erential operator L becomes L = 12 dX i @2 @x2i: When the spatial domain of the SPDE (5.3) is a subset of Rd and the boundary condition ut(x) = (t;x); on[0;T] @ is prescribed, the corresponding BDSDE is de ned through a stopping time de ned as = inffs;Xt;xs 2@ ;s t;x2 g: (5.6) Then, we have the BDSDE with stopping time as follows. Yt;xs = (Xx ^T) + Z T^ s f(r;Xt;xr ;Yr;xr ;Zt;xr )dr + Z T^ s g(r;Xt;xr ;Yr;xr ;Zt;xr )d Br Z T^ s Zt;xr dWr; t s T; x2 ; (5.7) where (Xx ;T) = h(Xt;xT )I T + ( ;Xt;x )I T. When t = 0, the stopping time de ned in (5:6) becomes = inffs;X0;xs 2@ ; s 0x2 g: Thus, BDSDE (5.7) changes to the following equation: Y 0;xt = (X0;x ^T) + Z T^ t f(s;X0;xs ;Y 0;xs ;Z0;xs )ds Z T^ t Z0;xs dWs + Z T^ t g(s;X0;xs ;Y 0;xs ;Z0;xs )d Bs; t2[0;T]; x2 ; 67 where for given x, X0;x0 = x, (X0;x ^T) = h(X0;xT )I T + ( ;X0;x )I T. The related SPDE is 8 >>>> >>< >>>> >>: ut(x) = h(x) + Z T t [12 dX i=1 @2us(x) @x2i +f(s;x;us(x);rus(x))]ds + Z T t g(s;x;us(x);rus(x))d Bs; x2 ; 0 t T; ut(x) = (t;x); on [0;T] @ : 5.2.1 Numerical algorithms For the simplicity of presentation we only consider the one dimensional case. The high dimensional cases can be handled through straightforward generalization of the one dimensional case. To simplify the notations we shall use (yt;zt) to denote the solution (Yt;xt ;Zt;xt ) of the BDSDE (5:2). We also denote Et;xs [X] = E[XjFW;t;xs ] where FW;t;xs := (x+Ws Wt;t s T)[ (Bt; 0 t T). Reference equations To further simplify the notations, we denotef(s;ys;zs) = f(s;Xt;xs ;ys;zs) andg(s;ys;zs) = g(s;Xt;xs ;ys;zs), knowing that x2 R. Then we have yt = yt+ + Z t+ t f(s;ys;zs)ds Z t+ t zsdWs + Z t+ t g(s;ys;zs) dBs; (5.8) where is a deterministic nonnegative number with t + T. Taking the conditional expectation Et;xt [ ] on (5.52), we obtain yt;xt = Et;xt [yt+ ] + Z t+ t Et;xt [f(s;ys;zs)]ds+ Z t+ t Et;xt [g(s;ys;zs)] dBs; (5.9) 68 where yt;xt = Et;xt [yt], that is, yt;xt is the value of yt at the time-space point (t;x). We use the simple right point formula to approximate the integrals in (5.53): Z t+ t Et;xt [f(s;ys;zs)]ds = Et;xt [f(t+ ;yt+ ;zt+ )] +RWy ; (5.10) Z t+ t Et;xt [g(s;ys;zs)] dBs = Et;xt [g(t+ ;yt+ ;zt+ )] Bt +RBy; (5.11) where RWy and RBy denote the corresponding errors of approximations. Inserting (5:54) and (5:11) into (5:53), we obtain yt;xt = Et;xt [yt+ ] + Et;xt [f(t+ ;yt+ ;zt+ )] +Et;xt [g(t+ ;yt+ ;zt+ )] Bt +Ry; (5.12) where Ry = RWy + RBy is the truncation error for solving yt. Let Ws = Ws Wt for t s t + . Multiplying by Wt+ on (5.52), taking the conditional expectation Et;xt [ ] and applying the It^o isometry we get Et;xt [yt+ Wt+ ] = Z t+ t Et;xt [f(s;ys;zs) Ws]ds + Z t+ t Et;xt [g(s;ys;zs) Ws] dBs Z t+ t Et;xt [zs]ds: (5.13) Similar to (5.12) we approximate the integrals in (5.13) with the right point formula to obtain Z t+ t Et;xt [f(s;ys;zs) Ws]ds = Et;xt [f(t+ ;yt+ ;zt+ ) Wt+ ] +RWz1 (5.14) Z t+ t Et;xt [zs]ds = zt;xt +RWz2 (5.15) and Z t+ t Et;xt [g(s;ys;zs) Ws] dBs = Et;xt [g(t+ ;yt+ ;zt+ ) Wt+ ] Bt +RBz ; (5.16) 69 where zt;xt is the value of zt at the time-space point (t;x), and RWz1, RWz2 and RBz are the corresponding approximation errors. Inserting (5:14), (5:15) and (5:16) into (5:13), we get the second approximation equation for (5:52) as follows. Et;xt [yt+ Wt+ ] = Et;xt [f(t+ ;yt+ ;zt+ ) Wt+ ] zt;xt +Et;xt [g(t+ ;yt+ ;zt+ ) Wt+ ] Bt +Rz; (5.17) where Rz = RWz1 +RWz2 +RBz is the truncation error for solving zt. (5:12) and (5:17) are two key equations of solving BDSDE (5:52) numerically. We refer them as reference equations. Discrete scheme To derive a numerical algorithm from the reference equations (5.12) and (5.17), we introduce the following time partition on [0;T]. Rth =ftijti2[0;T];ti 0 is a constant to be determined later. By the Lipschitz continuity of f, we have By = E[( tnEtn[en+1f ])2] L( tn)2(E[en+1y ]2 +E[en+1z ]2): (5.26) Similarly for Cy, using Cauchy?s inequality and Young?s inequality, we obtain Cy = E[2(Etn[en+1y ] + Btn+1Etn[en+1g ] +Rny)( tnEtn[en+1f ])] tn 1 2 E[Etn[en+1y ] + Btn+1Etn[en+1g ] +Rny]2 + tn 2E[en+1f ]2 tn 3 2 E[Etn[en+1y ]2 + tn(L1Etn[en+1y ]2 +L2Etn[en+1z ]2) + (Rny)2] + tnE[L 2Etn[en+1y ]2 +L 2Etn[en+1z ]2] tnC2E[en+1y ]2 + ( tn)2C3(E[en+1y ]2 +E[en+1z ]2) + tnL 2E[en+1z ]2 +C4( t)2; (5.27) where 2 > 0 is a constant to be determined later. Combining (5.24), (5.25), (5.26) and (5.27) together, we obtain E[eny]2 E[jEtn[en+1y ]j2] + tnE[(1 + 1)jEtn[en+1g ]j2] + tnL 2E[en+1z ]2 +K1 tnE[en+1y ]2 +K2( tn)2E[en+1z ]2 +K3( t)2: (5.28) For enz, we have the identity tnenz = Etn[en+1y Wtn+1] + tnEtn[en+1f Wtn+1] + Btn+1Etn[en+1g Wtn+1] +Rnz: (5.29) 77 Taking square on both sides of equation (5.29) and then taking expectation we obtain E[ tnenz]2 = E[Etn[en+1y Wtn+1] + Btn+1Etn[en+1g Wtn+1] +Rnz]2 +E[( tn)2Etn[en+1f Wtn+1]2] +2E[(Etn[en+1y Wtn+1] + Btn+1Etn[en+1g Wtn+1] +Rnz) ( tnEtn[en+1f Wtn+1])] = E[(Etn[en+1y Wtn+1])2 + ( Btn+1)2(Etn[en+1g Wtn+1])2 + (Rnz)2 +2(Etn[en+1y Wtn+1])(Etn[en+1g Wtn+1]) Btn+1 + 2(Etn[en+1y Wtn+1])Rnz +2Etn[en+1g Wtn+1] Btn+1Rnz] +E[( tn)2Etn[en+1f Wtn+1]2] +2E[(Etn[en+1y Wtn+1] + Btn+1Etn[en+1g Wtn+1] +Rnz) ( tnEtn[en+1f Wtn+1])] = Az +Bz +Cz; (5.30) where Az = E[(Etn[en+1y Wtn+1])2 + ( Btn+1)2(Etn[en+1g Wtn+1])2 + (Rnz)2 +2(Etn[en+1y Wtn+1])(Etn[en+1g Wtn+1]) Btn+1 + 2(Etn[en+1y Wtn+1])Rnz +2Etn[en+1g Wtn+1] Btn+1Rnz]; Bz = E[( tn)2Etn[en+1f Wtn+1]2] and Cz = 2E[(Etn[en+1y Wtn+1] + Btn+1Etn[en+1g Wtn+1] +Rnz) ( tnEtn[en+1f Wtn+1])]: For any FWt adapted process Xt we have (Etn[Xtn+1 Wtn+1])2 = (Etn[(Xtn+1 Etn[Xtn+1]) Wtn+1])2 Etn[(Xtn+1 Etn[Xtn+1])]2 tn = tn(Etn[(Xtn+1)2] jEtn[Xtn+1]j2): (5.31) 78 For Az, using (5.31), Cauchy?s inequality and Young?s inequality, we have Az = E[(Etn[en+1y Wtn+1])2 + ( Btn+1)2(Etn[en+1g Wtn+1])2 + (Rnz)2 +2(Etn[en+1y Wtn+1])(Etn[en+1g Wtn+1]) Btn+1 + 2(Etn[en+1y Wtn+1])Rnz +2Etn[en+1g Wtn+1] Btn+1Rnz] E[(Etn[en+1y Wtn+1])2 + ( Btn+1)2(Etn[en+1g Wtn+1])2 + (Rnz)2 + (Etn[en+1y Wtn+1])2 + 1 (Rnz)2 + 1(Etn[en+1g Wtn+1] Btn+1)2 + 1 1 (Rnz)2] tnE[Etn[en+1y ]2 jEtn[en+1y ]j2] +( tn)2E[Etn[en+1g ]2 jEtn[en+1g ]j2] +E[(Rnz)2] + tnE[Etn[en+1y ]2 jEtn[en+1y ]j2] + 1 E[(Rnz)2] + 1( tn)2E[Etn[en+1g ]2 jEtn[en+1g ]j2] + 1 1 E[(Rnz)2] (1 + ) tnE[Etn[en+1y ]2 jEtn[en+1y ]j2] +(1 + 1)( tn)2E[Etn[en+1g ]2 jEtn[en+1g ]j2] +C5( t)3: (5.32) Under the conditions in the theorem, we have Bz = E[( tn)2Etn[en+1f Wtn+1]2] C6( t)3: (5.33) Similarly, using Cauchy?s inequality and Young?s inequality, we obtain Cz = 2E[(Etn[en+1y Wtn+1] + Btn+1Etn[en+1g Wtn+1] +Rnz) ( tnEtn[en+1f Wtn+1])] tn 1 2 E[(Etn[en+1y Wtn+1] + Btn+1Etn[en+1g Wtn+1] +Rnz)]2 + tn 2Etn[en+1f Wtn+1]2 ( tn)2 3 2 (E[en+1y ]2 + tnL1E[en+1y ]2 + tnL2E[en+1z ]2 + (Rnz)2) 79 +( tn)2L 2(E[en+1y ]2 +E[en+1z ]2) C7 ( tn)2E[en+1y ]2 + ( tn)3(E[en+1y ]2 +E[en+1z ]2) +( tn)2L 2E[en+1z ]2 +C8( t)3: (5.34) Here 1 and 2 are same as in equation (5.91), and is a positive constant which will be determined later. Combining (5.30), (5.32), (5.33) and (5.34) together, we get E[ tnenz]2 (1 + ) tnE[Etn[en+1y ]2 jEtn[en+1y ]j2] +(1 + 1)( tn)2E[Etn[en+1g ]2 jEtn[en+1g ]j2] +( tn)2L 2E[en+1z ]2 +K4( tn)2Etn[en+1y ]2 +K5( tn)3Etn[en+1z ]2 +K6( t)3: (5.35) Next we divide by tn(1 + ) on both sides of (5.35) to obtain tn 1 + E[e n z] 2 E[Et n[e n+1 y ] 2 jEt n[e n+1 y ]j 2] +( tn)1 + 11 + E[Etn[en+1g ]2 jEtn[en+1g ]j2] +( tn)L 21 + E[en+1z ]2 +K4 tnEtn[en+1y ]2 +K5( tn)2Etn[en+1z ]2 +K6( t)2: (5.36) Adding (5.36) to (5.91), we obtain E[eny]2 + tn1 + E[enz]2 E[jEtn[en+1y ]j2] + tnE[(1 + 1)jEtn[en+1g ]j2] + tnL 2E[en+1z ]2 +K1 tnE[en+1y ]2 +K2( tn)2E[en+1z ]2 +K3(h2 +E[ Z tn+1 tn jzs ztnj2]ds) +E[Etn[en+1y ]2 jEtn[en+1y ]j2] +( tn)1 + 11 + E[Etn[en+1g ]2 jEtn[en+1g ]j2] 80 +( tn)L 21 + E[en+1z ]2 +K4 tnEtn[en+1y ]2 +K5( tn)2Etn[en+1z ]2 +K6(( t)2 +E[ Z tn+1 tn jzs ztnj2]ds) = E[eny]2 + ( tn)1 + 11 + E[Etn[en+1g ]2] +( tn)1 + 11 + E[jEtn[en+1g ]j2] +( tn)L 2 + 2 + 21 + E[en+1z ]2 +G1 tnE[en+1y ]2 +G2( tn)2E[en+1z ]2 +G3( t)2 E[eny]2 + ( tn)(L2 1 + 11 + +L2 1 + 11 + +L 2 + 2 + 21 + )E[en+1z ]2 +G1 tnE[en+1y ]2 +G2( tn)2E[en+1z ]2 +G3( t)2: (5.37) Now we choose , 1 and 2, all positive, su ciently small such that L2(1 + 1) +L2 (1 + 1) +L(2 2 + 2) 1: This is possible since L2 < 1. Thus, by equation (5.37), we have E[eny]2 + tn1 + E[enz]2 E[eny]2 + tn1 + E[en+1z ]2 +T1 tn(E[en+1y ]2 + tn1 + E[en+1z ]2) +T2( t)2: Denote en := E[eny]2 + tn1+ E[enz]2. Then the above equation becomes en (1 +T1 t)en+1 +T2( t)2: 81 By Gronwall?s inequality, we have max 0 n N 1 (E[eny]2 + tn1 + E[enz]2) C t as required. Error estimate for z We rst construct an approximate solution ( ~Y; ~Z) with step process as follows. Let ~Yt n+1 = y n+1 + tn fn+1 + Bt n g n+1 andGt =FWt _FBT . By an extension of It^o?s martingale representation theorem, we can nd an Gt adapted process ~Zt, such that ~Yt n+1 = E[ ~Yt n+1jGtn] + Z tn+1 tn ~ZrdWr: (5.38) De ne a continuous approximate process ( ~Y; ~Z) as follows. ~Yt = yn+1 +fn+1 (tn+1 t) +gn+1 (Bt n+1 Bt) Z tn+1 t ~ZrdWr; t2(tn;tn+1];n = 0; N 1 (5.39) where fn+1 = f(tn+1;yn+1;zn+1) and gn+1 = g(tn+1;yn+1;zn+1): By (5.19) and (5.38) it?s easy to see that tnzn = Z tn+1 tn Etn[ ~Zr]dr: 82 Thus Z tn+1 tn E[(zs zn)2]ds = Z tn+1 tn E[(zs 1 t n Z tn+1 tn Etn[ ~Zr]dr)2]ds = Z tn+1 tn E[( 1 t n Z tn+1 tn Etn[zs ~Zr]dr)2]ds E Z tn+1 tn 1 tn Z tn+1 tn Etn[(zs ~Zr)2]drds 2( Z tn+1 tn E[Etn[(zr ~Zr)2]]dr + Z tn+1 tn 1 tn Z tn+1 tn E[Etn[(zs zr)2]]drds) 2( Z tn+1 tn E[(zr ~Zr)2]dr + ( tn)2): (5.40) Now we are ready to prove an error estimate for z. Theorem 5.2 Let (yt;zt) be the exact solution and (yn;zn) be the solution of the scheme (5:18) and (5:19). Assume that Hypothesis (5:20) holds and derivatives f0x, f0y, f0z, g0x, g0y and g0z of f and g are all bounded. Then for t su ciently small, we have N 1X n=0 E Z tn+1 tn (zs zn)2ds C t: Proof: For t2[tn;tn+1], let ety = yt ~Yt, etz = zt ~Zt, ft = f(t;yt;zt) and gt = g(t;yt;zt). Subtracting BDSDE (5.39) from (5.2) we have that ety = etn+1y + Z tn+1 t (fs fn+1)ds+ Z tn+1 t (gs gn+1)d Bs Z tn+1 t eszdWs: (5.41) Taking square on both sides of (5.41), applying It^o?s formula (Pardoux and Peng (1994)) and taking expectation, we have E[(ety)2] +E Z tn+1 t (esz)2ds = E[(etn+1y )2] + 2E Z tn+1 t esy (fs fn+1)ds+E Z tn+1 t (gs gn+1)2ds E[(etn+1y )2] + 1 0 E Z tn+1 t (esy)2ds+ 0E Z tn+1 t (fs fn+1)2ds+E Z tn+1 t (gs gn+1)2ds 83 E[(etn+1y )2] + 1 0 E Z tn+1 t (esy)2ds+E Z tn+1 t 2c(ys yn+1)2 + ( 1 + )(zs zn+1)2ds E[(etn+1y )2] + 1 0 E Z tn+1 t (esy)2ds+CE Z tn+1 t (ys ytn+1)2 + (ytn+1 yn+1)2ds +E Z tn+1 t ( 1 + 1 2 + )[(zs ( tn) 1 Z tn+2 tn+1 Etn+1[zr]dr)2] +( 1 + 2 + )[(( tn) 1 Z tn+2 tn+1 Etn+1[zr]dr zn+1)2]ds; (5.42) where 0, 1 and 2 are positive constants to be determined later. Since E[(ys yt)2]+E[(zs zt)2] Cjs tj, we have E Z tn+1 t (ys ytn+1)2ds C( t)2 and E Z tn+1 t (zs ( tn) 1 Z tn+2 tn+1 Etn+1[zr]dr)2ds = E Z tn+1 t (zs ztn+1 +ztn+1 ( tn) 1 Z tn+2 tn+1 Etn+1[zr]dr)2ds = 2E Z tn+1 t (zs ztn+1)2 + (( tn) 1 Z tn+2 tn+1 Etn+1[ztn+1 zr]dr)2ds C( t)2: Also, since tzn+1 = Z tn+2 tn+1 Etn+1[ ~Zr]dr, we have (( tn) 1 Z tn+2 tn+1 Etn+1[zr]dr zn+1)2 ( tn) 1 Z tn+2 tn+1 Etn+1[(zr ~Zr)2]dr: Thus we can rewrite (5.42) as E[(ety)2] +E Z tn+1 t (esz)2ds E[(etn+1y )2] +CE Z tn+1 t (etn+1y )ds +( 1 + 2 + )E Z tn+2 tn+1 (esz)2ds+C( t)2 + 1 0 E Z tn+1 t (esy)2ds: 84 Choose 1 and 2 small enough such that ( 1 + 2 + ) = K < 1 (since < 1). Then E[(ety)2] +E Z tn+1 t (esz)2ds C1E Z tn+1 t (esy)2ds+E[(etn+1y )2] +C2E Z tn+1 tn (etn+1y )2ds +KE Z tn+2 tn+1 (esz)2ds+C( t)2 (1 +C2 t)E[(etn+1y )2] +KE Z tn+2 tn+1 (esz)2ds+C( t)2 +C1E Z tn+1 t (esy)2ds: (5.43) By Gronwall?s inequality, we get E[(esy)2] C((1 +C2 t)E[(etn+1y )2] +E Z tn+2 tn+1 (esz)2ds+C( t)2) (5.44) for s2[tn;tn+1]. Now we let t = tn in (5.43) and substitute (5.44) in (5.43) to obtain E[(etny )2] +E Z tn+1 tn (esz)2ds (1 +C1 t)E[(etn+1y )2] +(K +C2 t)E Z tn+2 tn+1 (esz)2ds+C( t)2: Now by Theorem 1, we easily obtain E[(etny )2] +E Z tn+1 tn (esz)2ds E[(etn+1y )2] + (K +C t)E Z tn+2 tn+1 (esz)2ds+C( t)2: Let t be su ciently small such that C t + K L< 1, where L is a constant. Summing the above equation from n = 0 to n = N 1, we obtain (1 L) N 1X n=0 E Z tn+1 tn (esz)2ds C t+L Z tN tN 1 E[(esz)2]ds C t: (5.45) Through a similar argument, it?s easy to obtain Z tN tN 1 E[(esz)2]ds C t: (5.46) 85 By (5.40), (5.45) and (5.46), we conclude that N 1X n=0 E Z tn+1 tn (zs zn)2ds C t as required in Theorem 2. 5.2.3 Numerical experiments In this section we carry out numerical experiments to verify the rate of convergence results obtained in Section 3 and compare our numerical method with the nite di er- ence method for stochastic parabolic partial di erential equations ([?]). The conditional expectations in (5.18) and (5.19) can be evaluated using Monte Carlo method or Gaussian quadratures ([?, 90]. In our examples, we use the binomial tree method which is amount to two point Gaussian quadrature ([?]). Example 1: In the rst example, we consider the initial boundary value problem. ut(x) = exp(x T) sin(B(T)2 ) + Z T t [12 @ 2 @x2us(x) (x+ 1 8)us(x) s 2 @ @xus(x)]ds + Z T t 1 2 exp(x s) cos(B(T) B(s) 2 )d B s; ut( 1) = exp(( 1) t) sin(B(T) B(T)2 ); ut(1) = exp(1 t) sin(B(T) B(t)2 ): (5.47) 86 Table 5.1: Example 1 of Section 6.2 J NT (FD) error(FD) NT (BDSDE) error (BDSDE) 22 25 0:0213 24 0:200 23 27 0:0177 26 0:0383 24 29 0:0112 28 0:0106 25 211 0:00587 210 0:00376 26 213 0:00313 212 0:00152 We construct the SPDE (5.47) in such a way that ut(x) = exp(x t) sin(B(T) B(t)2 ) is the exact solution. The corresponding BDSDE is given by y0;x0 = exp(X0;xT T) sin(B(T)2 )I T + exp(X0;x ) sin(B(T) B( )2 )I T + Z T^ 0 [ (X0;xt + 18)y0;xt t2z0;xt ]ds Z T^ 0 z0;xt dWt + Z T^ 0 1 2 exp(X 0;x t t) cos(B(T) B(t) 2 )d B t: The numerical results are shown in Table 1 and Figure 1. Here J denotes the number of spatial partition grids, NT(FD) the number of time steps used in nite di erence method, NT(BDSDE) the number of time steps used in our method for solving the related BDSDE, and error(FD) and error(BDSDE) the errors of nite di erence method and our method, respectively. The results indicate that our algorithm is comparable to the the algorithm of solving the SPDE directly using the nite di erence scheme, with a little higher rate of convergence. Example 2: In this example, we consider the unbounded SPDE initial value problem. ut(x) = sin(x+T) cos(2BT) + Z T t [12 @ 2 @x2us(x) @ @xus(x)]ds + Z T t sin(W(s) +s)(sin(BT +Bs) cos(BT +Bs)) +us(x)d Bs; (5.48) 87 ?5 ?4.5 ?4 ?3.5 ?3 ?2.5 ?2?6.5 ?6 ?5.5 ?5 ?4.5 ?4 ?3.5 ?3 ?2.5 ?2 ?1.5 log(h) log(error) BDSDE FD 2 1 Figure 5.1: Example 1: Convergence comparison between the direct nite di erence scheme and our scheme Table 5.2: Example 2 of Section 6.2 J NT error(Y) (u) error(Z) (ru) 23 23 6:4096E 002 0:1188 24 24 3:2019E 002 9:0028E 002 25 25 1:4426E 002 6:3314E 002 26 26 7:2577E 003 4:1337E 002 27 27 3:5995E 003 2:7149E 002 where ut(x) = sin(x+t) cos(BT +Bt) is the solution of the SPDE (5:48). The corresponding FBDSDE is given by y0;x0 = sin(W(T) +T) cos(2BT) Z T 0 z0;xs ds + Z t 0 [sin(W(s) +s)(sin(BT +Bs) cos(BT +Bs))+0;xs ]d Bs Z t 0 z0;xs dWs: The errors are shown in Table 2, in which error(Y) and error(Z) are errors for Y and Z at time-space point (t;x) = (0;0), respectively. These data also con rm our rate of converence results. 88 ?5 ?4.5 ?4 ?3.5 ?3 ?2.5 ?2?6 ?5.5 ?5 ?4.5 ?4 ?3.5 ?3 ?2.5 ?2 log(h) log(error) ErrorY ErrorZ 2 1 Figure 5.2: Example 2: Convergence comparison between the approximations of y and z 5.3 First Order Numerical Algorithms To develop rst order numerical algorithms for BDSDEs, we simplify the BDSDEs system to Yt;xs = h(Wt;xT ) + Z T s f(r;Wt;xr ;Yt;xr ;Zt;xr )dr + Z T s g(r;Wt;xr ;Yt;xr )d Br Z T s Zt;xr dWr; s t T: (5.49) In addition to the de nitions and assumptions in subsection 5.1, we introduce variational equations of BDSDEs. 5.3.1 Variational Equations LetrYt;xs ,rZt;xs andrWt;xs be the variations of Yt;xs , Zt;xs and Wt;xs , respectively, with respect to x at time level s = t. Notice that that rWt;xs 1. To simplify the presentation, we write ys = Yt;xs , zs = Zt;xs , Ws = Wt;xs , rys =rYt;xs , rzs =rZt;xs and rWs =rWt;xs . From [61], f(ryt;rzt); 0 t Tg, is the unique solution of the equation ryt =ryT + Z T t rFsds+ Z T t rGsd Bs Z T t rzsdWs; (5.50) 89 where ryT = h0(XT)rXT; rFs = f0W(s;Ws;ys;zs) +f0y(s;Ws;ys;zs) rys +f0z(s;Ws;ys;zs) rzs; t s T and rGs = g0W(s;Ws;ys) +g0y(s;Ws;ys) rys; t s T: Here f0W, f0y, f0z and g0W, g0y are rst order partial derivatives of functions f and g with respect to Ws, ys and zs. By Proposition 2.3 of [61], zs =rys (rWs) 1 =rys; 0 s T: (5.51) 5.3.2 Reference equations In this subsection we approximate integrals in (5.49) with appropriate quadratures with rst order accuracy and name the resulting equations as reference equations (see [87]). For this purpose, we de ne FW;ts := (Wr;t r s)_ (Bp; 0 p T). Let E[X] denote the mathematical expectation of the random variable X and Et;xt [X] denote the conditional expectation E[XjFW;tt ] of the random variable X with Wt = x. To further simplify the notation, we denote f(s;ys;zs) := f(s;Ws;Bs;ys;zs); g(s;ys) := g(s;Ws;Bs;ys); g0t(s;ys) := g0t(s;Ws;Bs;ys); g0W(s;ys) := g0W(s;Ws;Bs;ys); g0B(s;ys) := g0B(s;Ws;Bs;ys); g0y(s;ys) := g0y(s;Ws;Bs;ys); g00WW(s;ys) := g00WW(s;Ws;Bs;ys); g00BB(s;ys) := g00BB(s;Ws;Bs;ys); and g00yy(s;ys) := g00yy(s;Ws;Bs;ys); where g0t, g0W, g0B, g0y and g0z are rst order partial deriva- tives with respect to variables t, Wt, Bt and yt, respectively; g00WW, g00BB and g00yy are second order partial derivatives with respect to variables Wt, Bt, and yt, respectively. 90 With the above notations and from equation (5.49), we have for n = 1; ;NT 1, ytn = ytn+1 + Z tn+1 tn f(s;ys;zs)ds Z tn+1 tn zsdWs + Z tn+1 tn g(s;ys) dBs: (5.52) Reference equation for ys We rst eliminate the forward It^o integral from(5.52). To this end we take the condi- tional expectation Etn;xtn [ ] on both sides of (5.52) to obtain ytn;xtn = Etn;xtn [ytn+1] + Z tn+1 tn Etn;xtn [f(s;ys;zs)]ds+ Z tn+1 tn Etn;xtn [g(s;ys;zs)] dBs; (5.53) where ytn;xtn = Etn;xtn [ytn] is the value of ytn at the space point x. Next we approximate the integrals in the above equation with appropriate quadrature formulas. For the rst integral on the right hand side of (5.53), we simply use right point formula to obtain Z tn+1 tn Etn;xtn [f(s;ys;zs)]ds = tnEtn;xtn [f(tn+1;ytn+1;ztn+1)] +RW;ny ; (5.54) where RW;ny = Z tn+1 tn fEtn;xtn [f(s;ys;zs)] Etn;xtn [f(tn+1;ytn+1;ztn+1)]gds: (5.55) For the backward It^o integral term in (5.53), the application of the right point formula will result in only a half order scheme (see [6]) and thus is inadequate for deriving a rst order algorithm. To obtain a rst order numerical scheme, we introduce It^o?s formula for doubly stochastic integrals ([61] Lemma 1.3). Let t = (t;Wt;Bt;yt), t = (1;0;0; f(t;yt;zt)), t = (0;0;1; g(t;yt)) and t = (0;1;0;zt). In light of (5.49), (5.50) and the identity (5.51), we have following SDE for t: t = s + Z t s rdr + Z t s rd Br + Z t s rdWr; 0 s 0, it follows from Young?s inequality that 2E[Etn;xtn [en+1y Wtn] Rnz] E[(Etn;xtn [en+1y Wtn])2] + 1 E[(Rnz)2]: (5.97) 105 Inserting (5.95), (5.96) and (5.97) into the equation (5.94) and using (5.93), we obtain the estimate ( tn)2E[(enz)2] (1 + )E[(Etn;xtn [en+1y Wtn])2] +C 2( tn)2 E[Etn;xtn [(en+1y )2]] + 2( tn)2 E[Etn;xtn [(en+1z )2]] + 3C3( tn)3 E[Etn;xtn [(en+1z )2]] +C E[(Rnz)2]; (5.98) where C 2, C are constants depending on 2, respectively and functions f and g. Dividing both sides of equation (5.98) by ( tn)(1 + ) and noticing that (Etn;xtn [en+1y Wtn])2 tn (Etn;xtn [(en+1y )2] (Etn;xtn [en+1y ])2); we obtain ( tn1 + )E[(enz)2] E[Etn;xtn [(en+1y )2] (Etn;xtn [en+1y ])2] +C 2 tn1 + E[Etn;xtn [(en+1y )2]] + 2 tn1 + E[Etn;xtn [(en+1z )2]] + 3C31 + ( tn)2 E[Etn;xtn [(en+1z )2]] +C ( tn) 1 11 + E[(Rnz)2]: Since 11 + < 1, the above inequality can be rewritten as ( tn1 + )E[(enz)2] E[Etn;xtn [(en+1y )2] (Etn;xtn [en+1y ])2] +C 2 tn E[Etn;xtn [(en+1y )2]] + 2 tn E[Etn;xtn [(en+1z )2]] + 3C3( tn)2 E[Etn;xtn [(en+1z )2]] +C ( tn) 1 E[(Rnz)2]: (5.99) 106 With (5.91) and (5.99) in hand, we are ready to drive the result of the theorem. First we add (5.99) to (5.91) to obtain E[(eny)2] + tn1 + E[(enz)2] E[Etn;xtn [(en+1y )2]] + (C 1 +C 2) tn E[Etn;xtn [(en+1y )2]] + ( 1 + 2) tn E[Etn;xtn [(en+1z )2]] + 3(C1 +C3)( tn)2 E[Etn;xtn [(en+1z )2]] + 3E[(Rny)2] + (E[R W;n y ]) 2 tn +C ( tn) 1 E[(Rn z) 2]: (5.100) For a xed positive constant , we choose positive numbers 1, 2 so that 1 + 2 11 + : Then, the application of Jensen?s inequality leads to E[(eny)2] + tn1 + E[(enz)2] (1 +C 1; 2 tn) (E[(en+1y )2] + tn1 + E[(en+1z )2]) + 3E[(Rny)2] + (E[R W;n y ]) 2 tn +C ( tn) 1 E[(Rn z) 2]; (5.101) where C 1; 2 is a constant depending on 1, 2 and functions f and g. Finally using the discrete Gronwall inequality, we obtain max 0 n NT 1 (E[(eny)2] + tn1 + E[(enz)2]) C (E[(eNTy )2] + tNT 11 + E[(eNTz )2]) + NT 1X n=0 f3E[(Rny)2] + (E[R W;n y ]) 2 tn +C ( tn) 1 E[(Rn z) 2]g (5.102) as required. We now apply Theorem 5.3 to obtain error estimates for the proposed numerical scheme (5.80) and (5.81). In light of (5.102), it su ces to estimate the truncation errors Rny, Rnz and RW;ny . First we need a couple of regularity results for the exact solution (yt;zt). 107 Lemma 6 [61] For bounded f, g and h, E[(yt;xs yt;xt )2] +E[ Z t s (zr)2dr] Cjt sj; (5.103) and for bounded function with bounded second-order derivatives, (E[ (t;yt;zt) (s;ys;zs)])2 C(t s)2 where C is a constant independent of s and t. Lemma 7 [61] Assume that f, g, h, f0W, f0y, f0z, g0W, g0y, g0z, and h0 are all bounded, then E[(zt;xs zt;xt )2] +E[ Z t s (rzr)2dr] Cjt sj: (5.104) where C is a constant independent of s and t. Now we are ready to derive estimates for the truncation errors RW;ny Rny and Rnz. Proposition 2 Assume that that functions f, g, g0B, g0W, g0y are Lipschitz continuous, and f, g and h are bounded. Furthermore, assume that for any s 2 [0;T], (x;y;z) ! (f(s;x;y;z);g(s;x;y)) of class C3, all derivatives are bounded on [0;T] Rd Rk and h2C3(Rd;Rk). Then we have the following estimates (i): (E[RW;ny ])2 C( tn)4; (5.105) (ii): E[(Rny)2] C( tn)3; (5.106) (iii): E[(Rnz)2] C( tn)4; (5.107) where C is a constant only related to functions f and g. 108 Proof: (i). By Lemma 6, the de nition of RW;ny given by (5.55), and the assumptions in the Proposition we have that (E[RW;ny ])2 = ( Z tn+1 tn E[f(s;ys;zs) f(tn+1;ytn+1;ztn+1)]ds)2 tn Z tn+1 tn (E[f(s;ys;zs) f(tn+1;ytn+1;ztn+1)])2ds C( tn)4; which is (5.105). (ii). By de nition (5.62), Rny = RB;ny +RW;ny . Thus it su ces to estimate RB;ny and RW;ny . For Rng(s) given by (5.56), it easy to see that E[(Rng1(s))2] = E[(Et;xt [Rng(s)])2] = O(( tn)2); (5.108) From the de nition of Rng2(s) given by (5.57) we have that E[(Rng2(s))2] = E[(Etn;xtn [ Z tn+1 s (g0B(r;yr) g0yg(r;yr))d Br (g0B(tn+1;ytn+1) g0y(tn+1;ytn+1) g(tn+1;ytn+1)) Z tn+1 s d Br])2] 2E[(Etn;xtn [ Z tn+1 s (g0B(r;yr) g0B(tn+1;ytn+1))d Br])2] + 2E[(Etn;xtn [ Z tn+1 s (g0y(r;yr) g(r;yr) g0y(tn+1;ytn+1) g(tn+1;ytn+1))d Br])2]: For the rst term on the right hand side of the above inequality, we have by Lemma 6, It^o?s isometry and the assumptions in the Proposition, that E[(Etn;xtn [ Z tn+1 s (g0B(r;yr) g0B(tn+1;ytn+1))d Br])2] = O(( tn)2); (5.109) 109 Similarly, E[(Etn;xtn [ Z tn+1 s (g0y(r;yr) g(r;yr) g0y(tn+1;ytn+1) g(tn+1;ytn+1))d Br])2] = O(( tn)2): (5.110) Thus E[(Rng2(s))2] O(( tn)2): (5.111) It follows from (5.108) and (5.111) that E[(RB;ny )2] = E[( Z tn+1 tn (Rng1(s) +Rng2(s))d Bs)2] = O(( tn)3): (5.112) For RW;ny , it follows from Lemma 6, Lemma 7 and the assumptions of the Proposition that E[(RW;ny )2] = E[( Z tn+1 tn fEtn;xtn [f(s;ys;zs)] Etn;xtn [f(tn+1;ytn+1;ztn+1)]gds)2] tn E[ Z tn+1 tn fEtn;xtn [f(s;ys;zs)] Etn;xtn [f(tn+1;ytn+1;ztn+1)]g2ds] (5.113) tn E[ Z tn+1 tn (C( tn)2 +LEtn;xtn [(ys ytn+1)2] +LEtn;xtn [(zs ztn+1)2])ds] = O(( tn)3): Combing (5.111) with (5.113), we obtain the desired estimate (5.106) for Rny. (iii). At last, we derive estimate (5.107) for Rnz, which is de ned by Rnz = Rs;nz +RW;nz + RB;nz in (5.72). To this end, we derive the estimates for Rs;nz , RW;nz and RB;nz . We rst estimate the error term Rs;nz . Under the assumptions in the Proposition, E[ Z tn+1 tn (rFs)2ds] = O( tn) and E[(Etn;xtn [rGr] Etn;xtn [rGtn+1])2] = O( tn): 110 It follows from the de nition of Rs;nz given by (5.66) and the above two identities that E[(Rs;nz )2] = E[( Z tn+1 tn Z s tn Etn;xtn [rFr]drds Z tn+1 tn Z s tn (Etn;xtn [rGr] Etn;xtn [rGtn+1])d Brds)2] 2E[( Z tn+1 tn Z s tn Etn;xtn [rFr]drds)2] (5.114) +2E[( Z tn+1 tn Z s tn (Etn;xtn [rGr] Etn;xtn [rGtn+1])d Brds)2] = O(( tn)4): Similarly to (5.113), E[(RW;nz )2] = E[( Z tn+1 tn Etn;xtn [(f(s;ys;zs) f(tn+1;ytn+1;ztn+1)) Wtn]ds)2] = O(( tn)4): (5.115) To get the estimate RB;nz given by ( 5.70 ), we need the estimates for RB;nz1 , RB;nz2 and RB;nz3 . With estimate (5.108), (5.109) and (5.110) in hand, it?s easy to obtain E[(Et;xt [Rng(s) Wtn])2] = O(( tn)3); E[(Etn;xtn [( Z tn+1 s (g0B(r;yr) g0B(tn+1;ytn+1))d Br) Wtn])2] = O(( tn)3); (5.116) and E[(Etn;xtn [( Z tn+1 s g0 y(r;yr) g(r;yr) g 0 y(tn+1;ytn+1) g(tn+1;ytn+1) d B r) Wtn])2] = O(( tn)3): (5.117) Thus, we get the estimates for RB;nz1 and RB;nz2 as following E[(RB;nz1 (s))2] = E[(Et;xt [Rng(s) Wtn])2] = O(( tn)3);p 111 and E[(RB;nz2 (s))2] = E[( Z tn+1 s Etn;x tn g0 B(r;yr) g 0 y(r;yr) g(r;yr) W tn Etn;xtn g0B(tn+1;ytn+1) g0y(tn+1;ytn+1) g(tn+1;ytn+1) Wtn d Br)2] 2E[(Etn;xtn [( Z tn+1 s (g0B(r;yr) g0B(tn+1;ytn+1))d Br) Wtn])2] +2E[(Etn;xtn [( Z tn+1 s (g0y(r;yr) g(r;yr) g0y(tn+1;ytn+1) g(tn+1;ytn+1))d Br) Wtn])2] = O(( tn)3): By using the following estimates obtained directly from Lemma 6, Lemma 7 and the assumptions in the Proposition E[(Etn;xtn [ Z tn+1 s (g0W(r;yr) g0W(tn+1;ytn+1))dr])2] = O(( tn)3); (5.118) E[(Etn;xtn [ Z tn+1 s (g0y(r;yr) zr g0y(tn+1;ytn+1) ztn+1)dr])2] = O(( tn)3): (5.119) We get an estimate for RB;nz3 E[(RB;nz3 (s))2] = E[( Z tn+1 s Etn;x tn [g 0 W(r;yr) +g 0 y(r;yr) zr Etn;xtn g0W(tn+1;ytn+1) +g0y(tn+1;ytn+1) ztn+1 dr)2] 2E[(Etn;xtn [ Z tn+1 s (g0W(r;yr) g0W(tn+1;ytn+1))dr])2] +2E[(Etn;xtn [ Z tn+1 s (g0y(r;yr) zr g0y(tn+1;ytn+1) ztn+1)dr])2] = O(( tn)3): 112 Therefore, E[(RB;nz )2] = E[ Z tn+1 tn (RB;nz1 (s) +RB;nz2 (s) +RB;nz3 (s))d Bs 2 ] = O(( tn)4): (5.120) Then, from estimate (5.114), (5.115) and (5.120), we have E[(Rnz)2] C( tn)4: (5.121) Combining Theorem 5.3 and Proposition 2, we obtain the error estimates for our nu- merical scheme (5.80), (5.81). Theorem 5.4 Under the conditions of Theorem 5.3 and Proposition 2, if yNT = ytNT and zNT = ztNT , we have max 0 n NT 1 (E[(eny)2]) C( t)2; max 0 n NT 1 (E[(enz)2]) C t: (5.122) 5.3.5 Fully discrete scheme and its error estimate Fully discrete scheme The scheme we provided above is a semi-discrete scheme. In order to solve for (yn;zn) numerically, spatial approximations are needed. To simplify the presentation, in this section, we only consider the one dimensional case. The multi-dimensional cases can be deduced through a straightforward generalization. In addition to the temporal partition provided in Section 3, we introduce the following spatial partition of the real line R: Rh =fxijxi2R;i2Z;xi 0 and functions f and g. Recall that for a given random variable , ^Etn;xit n [ ] = 1p KX j=1 wj (xi +p2 taj): (5.130) To present the error estimates for the fully discrete scheme, we need the following Lemma Lemma 8 We assume that u, v are two functions, u;v : R ! R, and is any random variable. Then the following inequalities hold (^Etn;xitn [^u ^v])2 max i (u(xi) v(xi))2; ^Etn;xit n [(^u ^v) 2] max i (u(xi) v(xi))2; (5.131) (^Etn;xitn [ Wtn])2 t h^ Etn;xitn [ 2] (^Etn;xitn [ ])2 i (5.132) and (^Etn;xitn [ Wtn])2 t ^Etn;xitn [ 2]: (5.133) 118 Proof: From the de nition of ^Etn;xitn [ ] and the fact that 1p KX j=1 wj = 1, (^Etn;xitn [^u ^v])2 = ( 1p KX j=1 wj h ^u(xi +p2 t aj) ^v(xi +p2 t aj) i )2 ( 1p KX j=1 wj ^u(xi + p2 t a j) ^v(xi + p2 t a j) )2 ( 1p KX j=1 wj max i ju(xi) v(xi)j)2 = max i ju(xi) v(xi)j 2 = max i (u(xi) v(xi))2 and ^Etn;xit n [(^u ^v) 2]) = 1p KX j=1 wj ^u(xi +p2 t aj) ^v(xi +p2 t aj) 2 1p KX j=1 wj max i (u(xi) v(xi))2 = max i (u(xi) v(xi))2: as required in (5.131). To prove (5.132), we set ij = (xi +p2 t aj), it follows from the Cauchy Schwarz inequality and the fact 1p KX j=1 p2 t a j wj = 0 that ^ Etn;xitn [ Wtn] 2 = ( 1p KX j=1 p2 t a j wj ij)2 = ( 1p KX j=1 p2 t a j wj [ ij KX p=1 wpp i p ]) 2 [ KX j=1 wjp ( p2 t a j)2 ] [ KX j=1 wjp ( i j KX p=1 wpp i p) 2 ] 119 Since PKj=1 2wjp (aj)2 = 1 and PKj=1 wjp = 1, we see that [ KX j=1 wjp ( p2 t a j)2 ] [ KX j=1 wjp ( i j KX p=1 wpp i p) 2 ] = t [ KX j=1 2wjp (aj) 2 ] [ KX j=1 wjp ( ( i j) 2 2 i j KX p=1 wpp i p + ( KX p=1 wpp i p) 2 ) ] = t [ KX j=1 wjp ( i j) 2 ( KX j=1 wjp i j) 2 ] = t [ ^Etn;xitn [ 2] (^Etn;xitn [ ])2 ]; which proves the inequality (5.132). The inequality (5.133) follows directly from (5.132). Similar to the proof of (5.131), ^ Etn;xitn [( ^f(tn+1; ^ytn+1;^ztn+1) ^f(tn+1; ^yn+1;^zn+1))] 2 = [ 1p KX j=1 wj ^ f(tn+1; ^ytn+1(xi +p2 t aj);^ztn+1(xi +p2 t aj)) ^f(tn+1; ^yn+1(xi +p2 t aj);^zn+1(xi +p2 t aj)) ]2 [ 1p KX j=1 wj max i f(tn+1;ytn+1;xitn+1 ;ztn+1;xitn+1 ) f(tn+1;yn+1i ;zn+1i ) ]2 = max i f(tn+1;ytn+1;xitn+1 ;ztn+1;xitn+1 ) f(tn+1;yn+1i ;zn+1i ) 2 L [max i (en+1;iy )2 + max i (en+1;iz )2]: Following the above procedure, we see that ^ Etn;xitn [(^g(tn+1; ^ytn+1) ^g(tn+1; ^yn+1))] 2 Lmax i (en+1;iy )2; ^ Etn;xitn [(^g0B(tn+1; ^ytn+1) ^g0B(tn+1; ^yn+1))] 2 Lmax i (en+1;iy )2; ^ Etn;xitn [(^g0B(tn+1; ^ytn+1)^g(tn+1; ^ytn+1) ^g0B(tn+1; ^yn+1)^g(tn+1; ^yn+1))] 2 Lmax i (en+1;iy )2: (5.134) 120 Proof of Theorem 5.5: Notice that for any random variable ?tn+1 and ?n+1, Etn;xitn [?tn+1] ^Etn;xitn [ ^?n+1] = Etn;xitn [?tn+1] ^Etn;xitn [?tn+1] + ^Etn;xitn [?tn+1 ^?tn+1] + ^Etn;xitn [ ^?tn+1 ^?n+1]: (5.135) From the equation (5.128), we have that en;iy = ^Etn;xitn [^ytn+1 ^yn+1] + t f^Etn;xitn [ ^f(tn+1; ^ytn+1;^ztn+1) ^f(tn+1; ^yn+1;^zn+1)] +12 ^Etn;xitn [^g0B(tn+1; ^ytn+1) ^g0B(tn+1; ^yn+1)] 12 ^Etn;xitn [^g0y(tn+1; ^ytn+1) ^g(tn+1; ^ytn+1) ^g0y(tn+1; ^yn+1) ^g(tn+1; ^yn+1)]g+ Btn f^Etn;xitn [^g(tn+1; ^ytn+1) ^g(tn+1; ^yn+1)] 12 ^Etn;xitn [^g0B(tn+1; ^ytn+1) ^g0B(tn+1; ^yn+1)] Btn (5.136) +12 ^Etn;xitn [^g0y(tn+1; ^ytn+1) ^g(tn+1; ^ytn+1) ^g0y(tn+1; ^yn+1) ^g(tn+1; ^yn+1)] Btng +Iy1 +Iy2 +Rn;iy ; where Iy1 = Etn;xitn [ytn+1] ^Etn;xitn [ytn+1] + t fEtn;xitn [f(tn+1;ytn+1;ztn+1)] ^Etn;xitn [f(tn+1;ytn+1;ztn+1)] + 12Etn;xitn [g0B(tn+1;ytn+1)] 12 ^Etn;xitn [g0B(tn+1;ytn+1)] 12Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1)] + 12 ^Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1)]g + Btn fEtn;xitn [g(tn+1;ytn+1)] ^Etn;xitn [g(tn+1;ytn+1)] 12Etn;xitn [g0B(tn+1;ytn+1)] Btn + 12 ^Etn;xitn [g0B(tn+1;ytn+1)] Btn +12Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1)] Btn 12 ^Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1)] Btng; and Iy2 = ^Etn;xitn [ytn+1 ^ytn+1] + t f^Etn;xitn [f(tn+1;ytn+1;ztn+1) ^f(tn+1; ^ytn+1;^ztn+1)] 121 +12 ^Etn;xitn [g0B(tn+1;ytn+1) ^g0B(tn+1; ^ytn+1)] 12 ^Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1) ^g0y(tn+1; ^ytn+1) ^g(tn+1; ^ytn+1)]g+ Btn f^Etn;xitn [g(tn+1;ytn+1) ^g(tn+1; ^ytn+1)] 12 ^Etn;xitn [g0B(tn+1;ytn+1) ^g0B(tn+1; ^ytn+1)] Btn +12 ^Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1) ^g0y(tn+1; ^ytn+1) ^g(tn+1; ^ytn+1)] Btng; Similarly, the equation (5.129) becomes ten;iz = ^Etn;xitn [(^ytn+1 ^yn+1) Wtn+1] + t f^Etn;xitn [( ^f(tn+1; ^ytn+1;^ztn+1) ^f(tn+1; ^yn+1;^zn+1)) Wtn+1] +12 ^Etn;xitn [(^g0B(tn+1; ^ytn+1) ^g0B(tn+1; ^yn+1)) Wtn+1] 12 ^Etn;xitn [(^g0y(tn+1; ^ytn+1) ^g(tn+1; ^ytn+1) ^g0y(tn+1; ^yn+1) ^g(tn+1; ^yn+1)) Wtn+1]g + Btn f^Etn;xitn [(^g(tn+1; ^ytn+1) ^g(tn+1; ^yn+1)) Wtn+1] 12 ^Etn;xitn [(^g0B(tn+1; ^ytn+1) ^g0B(tn+1; ^yn+1)) Wtn+1] Btn (5.137) +12 ^Etn;xitn [(^g0y(tn+1; ^ytn+1) ^g(tn+1; ^ytn+1) ^g0y(tn+1; ^yn+1) ^g(tn+1; ^yn+1)) Wtn+1] Btng+Iz1 +Iz2 +Rn;iz ; where Iz1 = Etn;xitn [ytn+1 Wtn+1] ^Etn;xitn [ytn+1 Wtn+1] + t fEtn;xitn [f(tn+1;ytn+1;ztn+1) Wtn+1] ^Etn;xitn [f(tn+1;ytn+1;ztn+1) Wtn+1] +12Etn;xitn [g0B(tn+1;ytn+1) Wtn+1] 12 ^Etn;xitn [g0B(tn+1;ytn+1) Wtn+1] 12Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1) Wtn+1] +12 ^Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1) Wtn+1]g + Btn fEtn;xitn [g(tn+1;ytn+1) Wtn+1] ^Etn;xitn [g(tn+1;ytn+1) Wtn+1] 12Etn;xitn [g0B(tn+1;ytn+1) Wtn+1] Btn + 12 ^Etn;xitn [g0B(tn+1;ytn+1) Wtn+1] Btn 122 +12Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1) Wtn+1] Btn 12 ^Etn;xitn [g0y(tn+1;ytn+1) g(tn+1;ytn+1) Wtn+1] Btng; Iz2 = ^Etn;xitn [(ytn+1 ^ytn+1) Wtn+1] + t f^Etn;xitn [(f(tn+1;ytn+1;ztn+1) ^f(tn+1; ^ytn+1;^ztn+1)) Wtn+1] +12 ^Etn;xitn [(g0B(tn+1;ytn+1) ^g0B(tn+1; ^ytn+1)) Wtn+1] 12 ^Etn;xitn [(g0y(tn+1;ytn+1) g(tn+1;ytn+1) ^g0y(tn+1; ^ytn+1) ^g(tn+1; ^ytn+1)) Wtn+1]g + Btn f^Etn;xitn [(g(tn+1;ytn+1) ^g(tn+1; ^ytn+1)) Wtn+1] 12 ^Etn;xitn [(g0B(tn+1;ytn+1) ^g0B(tn+1; ^ytn+1)) Wtn+1] Btn +12 ^Etn;xitn [(g0y(tn+1;ytn+1) g(tn+1;ytn+1) ^g0y(tn+1; ^ytn+1) ^g(tn+1; ^ytn+1)) Wtn+1] Btng; For any function ?, jEtn;xitn [?] ^Etn;xitn [?]j D? K!2K(2K)!; where D? is the upper bound for j?(2K)x j . Since y, z, f, g, g0B, g0y 2C(2K)b , we have E[(Iy1 ])2 C K! 2K(2K)! 2 and E[max i (en;iy )2] + t1 + E[max i (en;iz )2] (1 +C t) (E[max i (en+1;iy )2] + t1 + E[max i (en+1;iz )2]) +3E[(Iy1 +Iy2 + max i Rn;iy )2] + (E[I y 1 +I y 2 + maxiR W;n;i y ]) 2 t +C ( t) 1 E[(Iz1 +Iz2 + max i Rn;iz )2]: 123 Therefore, E[max i (en;iy )2] + t1 + E[max i (en;iz )2] (1 +C t) (E[max i (en+1;iy )2] + t1 + E[max i (en+1;iz )2]) +3C " K! 2K(2K)! 2 +h4 + ( t)3 # +C( t) 1 " K! 2K(2K)! 2 +h4 + ( t)4 # +C ( t) 1 C " K! 2K(2K)! 2 +h4 + ( t)4 # = (1 +C t) (E[max i (en+1;iy )2] + t1 + E[max i (en+1;iz )2]) +C ( t) 1 C " K! 2K(2K)! 2 +h4 + ( t)4 # : (5.138) By applying discrete Grownwall?s inequality, we obtain E[max i (en;iy )2] + t1 + E[max i (en;iz )2] NTX j=n C ( t) 1 C " K! 2K(2K)! 2 +h4 + ( t)4 # : Thus, max 0 n NT 1 E[max i (en;iy )2] C " ( t) 2 K! 2K(2K)! 2 + h 4 ( t)2 + ( t) 2 # and max 0 n NT 1 E[max i (en;iz )2] C " ( t) 3 K! 2K(2K)! 2 + h 4 ( t)3 + t # ; as required. Remark 3 Notice that the error terms ( t) 2 K! 2K(2K)! 2 and ( t) 3 K! 2K(2K)! 2 decrease very rapidly as the parameter K becomes large. 5.3.6 Numerical experiment We now demonstrate the e ectiveness and accuracy of our method for solving the BDS- DE (5.49). We will compare our rst order scheme with the half order scheme ( see [6] ) in 124 Table 5.3: Example 1 of Section 6.3 NT Error1 2 (Y) Error1 2 (Z) Error1(Y) Error1(Z) 24 0:2899 0:4197 0:1036 0:1486 25 0:2162 0:3193 5:35E 02 7:74E 02 26 0:13478 0:1942 2:52E 02 3:74E 02 27 8:76E 02 0:1292 1:21E 02 1:81E 02 CR 0:5864 0:5817 1:0380 1:0162 example1. In example 2, we show an application of BDSDE in solving a nonlinear ltering problem. Example 1. Consider the BDSDE Yt;xt = exp(T +Wt;xT ) + Z T t sin(s) Yt;x s Z t;x s sin(s) exp(s+W t;x s +Bs BT) ds Z T t Zt;xs dWs Z T t Yt;xs d Bs: The exact solution is Yt;xs = exp(s+Wt;xs +Bs BT), Zt;xs = exp(s+Wt;xs +Bs BT). Numerical results are shown in Table 5.3. Here, the integer NT is the number of temporal partitions, CR is the convergence rate, Error1(Y) and Error1(Z) represent, respectively, the errors in the approximation of Yt;xt and Zt;xt of the rst order scheme. Also, Error1 2 (Y) and Error1 2 (Z) are respectively, the errors in approximation of Yt;xt and Zt;xt of the half order scheme. The results verify the theoretical error estimates we obtained in Section 4. Example 2. We present a practical example which illustrates the application of our numerical method to solve a nonlinear ltering problem. This example is a classical \bearing- only tracking" problem (see [8]). The simulation scenario is shown in Figure 5.3. Consider a target moving along the x-axis, according to a state equation dXt = (4 + 2 sinXt)dt+d ~Wt; X0 = 40: 125 Figure 5.3: Target tracking by using one detector 0 5 10 15 20 25 30 35 40 45 500 5 10 15 20 25 30 Observation Angle v = 4 + 2sin x (m/s) Platform Target A xed observer located on a platform on the y-axis takes noisy measurement Ot of the target bearing dOt = tan 1( 20X t ) +d ~Bt; O0 = 0: Here, ~Wt and ~Bt are two independent standard Brownian Motions. The goal is to nd Figure 5.4: Tracking Estimate 0 2 4 6 8 10 120 5 10 15 20 25 30 35 40 45 Time (seconds) Position (meters) Real state Estimate state the best estimate of the target position Xt based on the observations up to time t, i.e. for E[XtjF(Ot0)], where F(Ot0) denotes the -algebra generated by the observation fOsg0 s t. The initial guess of the position of the target is X0 = 44. One has the conditional density 126 Figure 5.5: Probability Distribution (a) 30 steps 25 30 35 400 0.005 0.01 0.015 0.02 0.025 0.03 0.035 0.04 0.045 0.05 Position Probability density Distribution Real Target (b) 60 steps 10 15 20 25 300 0.01 0.02 0.03 0.04 0.05 0.06 Position Probability density Distribution Real Target (c) 90 steps 0 2 4 6 8 100 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Position Probability density Distribution Real Target p p(XtjF(Ot0)) = YT t;XtT t , where YT t;xs satis es the following BDSDE( see [61, ?]) YT t;xs = YT t;xT + Z T s 2 cos(WT t;x r ) Y T t;x r + (4 + 2 sin(W T t;x r )) Z T t;x r dr Z T s ZT t;xr dWr + Z T s tan 1( 20WT t;x r ) YT t;xr d OT r; where T t s T, and x 2 R. The initial condition YT t;xT of the above equation is the normal distribution N(44;1). The stochastic process Wt is a standard Brownian motion independent of the observation process Ot. De ne the equivalent probability measure Q by dQ dPjGt = exp Z t 0 tan 1( 20X t )dOs 12 Z t 0 jtan 1( 20X t )j2ds ; t2R+; where Gt is the -algebra generated by fXsg0 s t and fOsg0 s t, t2R+. One has under probability measure Q, that Wt and Ot are two independent standard Brownian Motions(see [?]). In our simulation example, we assume that the observations are collected at intervals of length t = 0:125s and we track this target for 12s. Figure 5.4 illustrates the comparison of simulation results and the real target state. Figure 5.5 shows distributions of the simulated conditional probability p(XtjF(Ot0)). We present the simulation results to three time levels: 127 t = 3:75s(30 t);7:5s(60 t);11:25s(90 t). The red lines represent the real target positions and the blue curves show the simulated distribution. 128 Chapter 6 Concluding Remarks The major contribution of this work is the development of three novel numerical ap- proximation methods for the solution of nonlinear ltering problems: (i) an implicit lter method; (ii) a hybrid sparse grid Zakai lter method; (iii) the FBDSDE lter method. Our rst e ort in this work is the development of an implicit lter algorithm for nonlinear ltering problems. This method is under the general framework of the Bayesian lter. The essential idea is to evaluate the probability distribution of the current state in the prediction step by evaluating the previous state through the state equation, given the value of the current state and a sample of the noise. Through rigorous analysis we proved the convergence of the algorithm. Numerical experiments indicate that our method is more accurate than the standard Kalman lter and is more stable than the standard particle lter method for long term simulations. It needs to point out that our method is a grid method in which the probability distributions are evaluated at all grid points. As such, the computing cost will increase exponentially as the dimension increases. Thus our method at current form is suitable for only low dimension problems such as target tracking problems.. In the future, we plan to improve the algorithm by adding an adaptive mechanism and a new interpolation method called the \radial basis approximations" to it so that it will be more e cient in solving higher dimensional problems. The second e ort in this work focuses on the Zakai lter. We proposed the construction of a hybrid numerical algorithm to improve the e ciency of the Zakai lter for moderately high dimensional nonlinear ltering problems. Our algorithm combines the advantages of the splitting-up approximation scheme for the Zakai lter, a hierarchical sparse grid method for moderately high dimensional problems to compute the numerical solution of the Zakai lter, 129 and an importance sampling method to adaptively construct a bounded domain at each time step of the temporal discretization. The hierarchical sparse grid method reduces the number of grid points we need to solve in the splitting-up approximation scheme for the Zakai lter. The application of the solution domain adaptive method allows us to solve the Zakai lter in the high density region of the target state pdf which reduces the size of domains that we build the sparse grid points. We used one numerical experiment to show the e ectiveness of our solution domain adaptive method and two other examples to demonstrate that our algorithm is more e cient than the standard Zakai lter while maintaining its high accuracy. In addition, we proposed two numerical schemes for computing backward doubly s- tochastic di erential equations (BDSDEs). Becasue of the equivalent relation between BDS- DEs and SPDEs, our schemes can also be used to nd numerical solutions of the Zakai equation. Thus our algorithms also provide numerical methods for solving nonlinear lter- ing problems. It is worthy to mention that when solving the BDSDEs, we only need to solve stochastic ordinary di erential equations instead of solving SPDEs in the Zakai lter. In this connection, high order numerical approximation methods for stochastic ordinary di erential equations can be applied to develop high order numerical schemes for BDSDEs. The rst numerical scheme we proposed in this work provides a half order numerical approximation method for BDSDEs which is derived by using the Euler scheme to discretize the stochastic integrals. The second numerical scheme gives a rst order numerical approximation method. The main idea of the rst order scheme is to use the two-sided It^o-Taylor expansion for forward and backward stochastic di erentials to construct high order quadratures for the stochastic integrals involving both backward and forward Brownian motions. Through rig- orous analysis, we proved the convergence rates for both the half order scheme and the rst order scheme. Although we only introduced a rst order numerical scheme in this work, higher order schemes can also be developed using higher order It^o Taylor expansions. 130 Bibliography [1] N. U. Ahmed and S. M. Radaideh. A powerful numerical technique solving Zakai equation for nonlinear ltering. Dynam. Control, 7(3):293{308, 1997. [2] K. J. Astr om, G. C. Goodwin, and P. R. Kumar, editors. Adaptive control, ltering, and signal processing, volume 74 of The IMA Volumes in Mathematics and its Applications. Springer-Verlag, New York, 1995. [3] Tuncer Can Aysal and Kenneth E. Barner. Meridian ltering for robust signal process- ing. IEEE Trans. Signal Process., 55(8):3949{3962, 2007. [4] Syed Baker, Hart Poskar, Falk Schreilber, and Bjorn Junker. An improved constraint ltering technique for inferring hidden states and parameters of a biological model. Bioinformatics, 29:1052{1059, 2013. [5] V. Bally. Approximation scheme for solutions of BSDE. In Backward stochastic dif- ferential equations (Paris, 1995{1996), volume 364 of Pitman Res. Notes Math. Ser., pages 177{191. Longman, Harlow, 1997. [6] Feng Bao, Yanzhao Cao, and Weidong Zhao. Numerical solutions for forward backward doubly stochastic di erential equations and zakai equations. International Journal for Uncertainty Quanti cation, 1(4):351{367, 2011. [7] Yaakov Bar-Shalom and Thomas E. Fortmann. Tracking and data association, volume 179 of Mathematics in Science and Engineering. Academic Press Inc., San Diego, CA, 1988. [8] Yaakov Bar-Shalom and Thomas E. Fortmann. Tracking and data association, volume 179 of Mathematics in Science and Engineering. Academic Press Inc., San Diego, CA, 1988. [9] Christian Bender and Jessica Steiner. A posteriori estimates for backward sdes. SIAM/ASA J. Uncertainty Quanti cation, 1(1):139{163, 2013. [10] A. Bensoussan, R. Glowinski, and A. R a scanu. Approximation of some stochastic d- i erential equations by the splitting up method. Appl. Math. Optim., 25(1):81{106, 1992. [11] Alain Bensoussan, Jussi Keppo, and Suresh P. Sethi. Optimal consumption and portfolio decisions with partially observed real prices. Math. Finance, 19(2):215{236, 2009. 131 [12] A. V. Bernshte n and A. P. Kuleshov. Optimal ltering of a random background in image processing problems. Problemy Peredachi Informatsii, 44(3):70{80, 2008. [13] Jose Bins, Leandro L. Dihl, and Claudio R. Jung. Target tracking using multiple patches and weighted vector median lters. J. Math. Imaging Vision, 45(3):293{307, 2013. [14] AlMuatazbellah M. Boker and Hassan K. Khalil. Nonlinear observers comprising high- gain observers and extended Kalman lters. Automatica J. IFAC, 49(12):3583{3590, 2013. [15] Miodrag Boli c, Petar M. Djuri c, and Sangjin Hong. Resampling algorithms and archi- tectures for distributed particle lters. IEEE Trans. Signal Process., 53(7):2442{2450, 2005. [16] Bruno Bouchard and Nizar Touzi. Discrete-time approximation and Monte-Carlo simulation of backward stochastic di erential equations. Stochastic Process. Appl., 111(2):175{206, 2004. [17] Hans-Joachim Bungartz and Michael Griebel. Sparse grids. Acta Numer., 13:147{269, 2004. [18] Zhe Chen. Bayesian ltering: from kalman lters to particle lters, and beyond. Un- published manuscript, 2011. [19] D. Chevance. Numerical methods for backward stochastic di erential equations. In Numerical methods in nance, Publ. Newton Inst., pages 232{244. Cambridge Univ. Press, Cambridge, 1997. [20] J. M. C. Clark, S. A. Robbiati, and R. B. Vinter. The shifted Rayleigh mixture lter for bearings-only tracking of maneuvering targets. IEEE Trans. Signal Process., 55(7, part 1):3218{3226, 2007. [21] D. Crisan and O. Obanubi. Particle lters with random resampling times. Stochastic Process. Appl., 122(4):1332{1368, 2012. [22] Dan Crisan. Exact rates of convergence for a branching particle approximation to the solution of the Zakai equation. Ann. Probab., 31(2):693{718, 2003. [23] Dan Crisan. Exact rates of convergence for a branching particle approximation to the solution of the Zakai equation. Ann. Probab., 31(2):693{718, 2003. [24] Dan Crisan and Jie Xiong. A central limit type theorem for a class of particle lters. Commun. Stoch. Anal., 1(1):103{122, 2007. [25] Crisan D. and Doucet A. A survey of convergence results on particle ltering methods for practitiners. IEEE Trans. Sig. Proc. [26] D. H. Dini, C. Jahanchahi, and D. P. Mandic. Kalman ltering for widely linear complex and quaternion valued bearings only tracking. IET Signal Process., 6(5):435{445, 2012. 132 [27] Jind rich Dun k, Miroslav Simandl, and Ond rej Straka. Unscented Kalman lter: aspects and adaptive setting of scaling parameter. IEEE Trans. Automat. Control, 57(9):2411{ 2416, 2012. [28] Garry A. Einicke. Asymptotic optimality of the minimum-variance xed-interval s- moother. IEEE Trans. Signal Process., 55(4):1543{1547, 2007. [29] Bertil Ekstrand. Some aspects on lter design for target tracking. J. Control Sci. Eng., pages Art. ID 870890, 15, 2012. [30] Patrick Florchinger and Fran cois Le Gland. Time-discretization of the Zakai equation for di usion processes observed in correlated noise. Stochastics Stochastics Rep., 35(4):233{ 256, 1991. [31] R udiger Frey and Thorsten Schmidt. Pricing corporate securities under noisy asset information. Math. Finance, 19(3):403{421, 2009. [32] R udiger Frey, Thorsten Schmidt, and Ling Xu. On Galerkin approximations for the Zakai equation with di usive and point process observations. SIAM J. Numer. Anal., 51(4):2036{2062, 2013. [33] Paul Frogerais, Jean-Jacques Bellanger, and Lot Senhadji. Various ways to com- pute the continuous-discrete extended Kalman lter. IEEE Trans. Automat. Control, 57(4):1000{1004, 2012. [34] Emmanuel Gobet and C eline Labart. Error expansion for the discretization of backward stochastic di erential equations. Stochastic Process. Appl., 117(7):803{829, 2007. [35] Emmanuel Gobet, Jean-Philippe Lemor, and Xavier Warin. A regression-based Monte Carlo method to solve backward stochastic di erential equations. Ann. Appl. Probab., 15(3):2172{2202, 2005. [36] Emmanuel Gobet, Gilles Pag es, Huy^en Pham, and Jacques Printems. Discretization and simulation of the Zakai equation. SIAM J. Numer. Anal., 44(6):2505{2538 (electronic), 2006. [37] H. Gong, J. Singh, and A. Thavaneswaran. Filtering and option pricing with transfor- mation. J. Appl. Statist. Sci., 17(3, [2009 on cover]):363{375, 2010. [38] N.J Gordon, D.J Salmond, and A.F.M. Smith. Novel approach to nonlinear/non- gaussian bayesian state estimation. IEE PROCEEDING-F, 140(2):107{113, 1993. [39] Y. Hu, G. Kallianpur, and J. Xiong. An approximation for the Zakai equation. Appl. Math. Optim., 45(1):23{44, 2002. [40] Je rey Humpherys, Preston Redd, and Jeremy West. A fresh look at the Kalman lter. SIAM Rev., 54(4):801{823, 2012. 133 [41] John D. Jakeman, Richard Archibald, and Dongbin Xiu. Characterization of disconti- nuities in high-dimensional stochastic problems on adaptive sparse grids. J. Comput. Phys., 230(10):3977{3997, 2011. [42] Adrew H. Jazwinski. Stochastic Processing and Filtering Theory, volume 64. Academic Press, New York, 1973. [43] Simon J. Julier and Joseph J. LaViola, Jr. On Kalman ltering with nonlinear equality constraints. IEEE Trans. Signal Process., 55(6, part 2):2774{2784, 2007. [44] Xiong Kai, Liu Liangdong, and Liu Yiwu. Robust extended Kalman ltering for non- linear systems with multiplicative noises. Optimal Control Appl. Methods, 32(1):47{63, 2011. [45] R. E. Kalman and R. S. Bucy. New results in linear ltering and prediction theory. Trans. ASME Ser. D. J. Basic Engrg., 83:95{108, 1961. [46] A. N. Kolmogoro . Stationary sequences in Hilbert?s space. Bolletin Moskovskogo Gosudarstvenogo Universiteta. Matematika, 2:40pp, 1941. [47] Gennady Yu. Kulikov and Maria V. Kulikova. Accurate numerical implementation of the continuous-discrete extended Kalman lter. IEEE Trans. Automat. Control, 59(1):273{ 279, 2014. [48] Fran cois Le Gland. Time discretization of nonlinear ltering equations. In Proceedings of the 28th IEEE Conference on Decision and Control, Vol. 1{3 (Tampa, FL, 1989), pages 2601{2606, New York, 1989. IEEE. [49] Binghuai Lin and Dennis McLaughlin. E cient Characterization of Uncertain Model Parameters with a Reduced-Order Ensemble Kalman Filter. SIAM J. Sci. Comput., 36(2):B198{B224, 2014. [50] G. Lin, A. M. Tartakovsky, and D. M. Tartakovsky. Uncertainty quanti cation via random domain decomposition and probabilistic collocation on sparse grids. J. Comput. Phys., 229(19):6995{7012, 2010. [51] J. Ma and J. Yong. Approximate solvability of forward-backward stochastic di erential equations. Appl. Math. Optim., 45(1):1{22, 2002. [52] Jin Ma, Philip Protter, Jaime San Mart n, and Soledad Torres. Numerical method for backward stochastic di erential equations. Ann. Appl. Probab., 12(1):302{316, 2002. [53] Jin Ma, Philip Protter, and Jiong Min Yong. Solving forward-backward stochastic di erential equations explicitly|a four step scheme. Probab. Theory Related Fields, 98(3):339{359, 1994. [54] Jin Ma, Jie Shen, and Yanhong Zhao. On numerical approximations of forward- backward stochastic di erential equations. SIAM J. Numer. Anal., 46(5):2636{2661, 2008. 134 [55] Xiang Ma and Nicholas Zabaras. An adaptive hierarchical sparse grid collocation algorithm for the solution of stochastic di erential equations. J. Comput. Phys., 228(8):3084{3113, 2009. [56] C. Johan Masreliez and R. Douglas Martin. Robust Bayesian estimation for the lin- ear model and robustifying the Kalman lter. IEEE Trans. Automatic Control, AC- 22(3):361{371, 1977. [57] Motoyasu Nagata and Yoshikazu Sawaragi. Error analysis of the Schmidt-Kalman lter. Internat. J. Systems Sci., 7(7):769{778, 1976. [58] Boris N. Oreshkin and Mark J. Coates. Analysis of error propagation in particle lters with approximation. Ann. Appl. Probab., 21(6):2343{2378, 2011. [59] Umut Orguner and Fredrik Gustafsson. Target tracking with particle lters under signal propagation delays. IEEE Trans. Signal Process., 59(6):2485{2495, 2011. [60] E. Pardoux and P. Protter. A two-sided stochastic integral and its calculus. Probab. Theory Related Fields, 76(1):15{49, 1987. [61] Etienne Pardoux and Shi Ge Peng. Backward doubly stochastic di erential equations and systems of quasilinear SPDEs. Probab. Theory Related Fields, 98(2):209{227, 1994. [62] Nigel George Peach. Bearings-only tracking using a set of range parameterised extended Kalman lters. ProQuest LLC, Ann Arbor, MI, 1997. Thesis (Ph.D.){University of Bath (United Kingdom). [63] A. Peirce and F. Rochinha. An integrated extended Kalman lter-implicit level set algorithm for monitoring planar hydraulic fractures. Inverse Problems, 28(1):015009, 22, 2012. [64] Dirk P uger. page 194, 2010. Thesis (Ph.D.){ Technische Universit at M unchen. [65] J. A. Pocock, S. L. Dance, and A. S. Lawless. State estimation using the particle lter with mode tracking. Comput. & Fluids, 46:392{397, 2011. [66] Charles F. Price. An analysis of the divergence problem in the Kalman lter. IEEE Trans. Automatic Control, AC-13:699{702, 1968. [67] Jacques Printems. On the discretization in time of parabolic stochastic partial di er- ential equations. M2AN Math. Model. Numer. Anal., 35(6):1055{1078, 2001. [68] Christoph Reisinger. Analysis of linear di erence schemes in the sparse grid combination technique. IMA J. Numer. Anal., 33(2):544{581, 2013. [69] Elliott Robet, J. and Kuen Tak Siu. Option pricing and ltering with hidden markov- modulated pure-jump processes. Appl. Math. Finance, 20(1):1{25, 2013. 135 [70] Jie Shen and Haijun Yu. E cient spectral sparse grid methods and applications to high-dimensional elliptic equations II. Unbounded domains. SIAM J. Sci. Comput., 34(2):A1141{A1164, 2012. [71] S. Singh, S. Digumarthy, A Back, J. Shepard, and M Kalra. Radiation dose reduction for chest ct with non-linear adaptive lters. Acta Radiologica, 55(2):169{174, 2013. [72] M. W. A. Smith and A. P. Roberts. An exact equivalence between the discrete- and continuous-time formulations of the Kalman lter. Math. Comput. Simulation, 20(2):102{109, 1978. [73] Subchan and Tahiyatul As hani. The missile guidance estimation using extended Kalman lter-unknown input-without direct feedthrough (EKF-UI-WDF) method. J. Indones. Math. Soc., 19(1):1{14, 2013. [74] Ali Akbar Tadaion, Mostafa Derakhtian, Saeed Gazor, and Mohammad R. Aref. A fast multiple-source detection and localization array signal processing algorithm using the spatial ltering and ML approach. IEEE Trans. Signal Process., 55(5, part 1):1815{ 1827, 2007. [75] Krystyna Twardowska, Tomasz Marnik, and Monika Pas lawaska-Po luniak. Approxi- mation of the Zakai equation in a nonlinear ltering problem with delay. Int. J. Appl. Math. Comput. Sci., 13(2):151{160, 2003. [76] Yener Ulker and Bilge Gunsel. Multiple model target tracking with variable rate particle lters. Digit. Signal Process., 22(3):417{429, 2012. [77] Xin Wang and Shu-Li Sun. Measurement feedback self-tuning weighted measurement fusion Kalman lter for systems with correlated noises. J. Appl. Math., pages 324296, 16, 2012. [78] Nick Whiteley and Anthony Lee. Twisted particle lters. Ann. Statist., 42(1):115{141, 2014. [79] N. Wiener and E. Hopf. On a class of singular integral equations. Proc. Prussian Acad. Math. - Phys. Ser., page 696pp, 1931. [80] Norbert Wiener. Extrapolation, Interpolation, and Smoothing of Stationary Time Series. With Engineering Applications. The Technology Press of the Massachusetts Institute of Technology, Cambridge, Mass, 1949. [81] Dongbin Xiu and Jan S. Hesthaven. High-order collocation methods for di erential equations with random inputs. SIAM J. Sci. Comput., 27(3):1118{1139 (electronic), 2005. [82] Moshe Zakai. On the optimal ltering of di usion processes. Z. Wahrscheinlichkeits- theorie und Verw. Gebiete, 11:230{243, 1969. 136 [83] Andreas Zeiser. Fast matrix-vector multiplication in the sparse-grid Galerkin method. J. Sci. Comput., 47(3):328{346, 2011. [84] Guannan Zhang and Max Gunzburger. Error analysis of a stochastic collocation method for parabolic partial di erential equations with random input data. SIAM J. Numer. Anal., 50(4):1922{1940, 2012. [85] H. Zhang and D. Laneuville. Grid based solution of zakai equation with adaptive local re nement for bearing-only tracking. IEEE Aerospace Conference, 2008. [86] Jianfeng Zhang. A numerical scheme for BSDEs. Ann. Appl. Probab., 14(1):459{488, 2004. [87] Weidong Zhao, Lifeng Chen, and Shige Peng. A new kind of accurate numerical method for backward stochastic di erential equations. SIAM J. Sci. Comput., 28(4):1563{1581, 2006. [88] WeiDong Zhao, Yang Li, and Yu Fu. Second-order schemes for solving decoupled forward backward stochastic di erential equations. Sci. China Math., 57(4):665{686, 2014. [89] Weidong Zhao, Yang Li, and Guannan Zhang. A generalized -scheme for solving back- ward stochastic di erential equations. Discrete Contin. Dyn. Syst. Ser. B, 17(5):1585{ 1603, 2012. [90] Weidong Zhao, Guannan Zhang, and Lili Ju. A stable multistep scheme for solving backward stochastic di erential equations. SIAM J. Numer. Anal., 48(4):1369{1394, 2010. [91] Y. Zhao, H. M. Guo, and S. A. Billings. Application of totalistic cellular automata for noise ltering in image processing. J. Cell. Autom., 7(3):207{221, 2012. 137