Abstract
Power penalty methods for solving a linear parabolic complementarity problem arising from American option pricing have attracted much attention. These methods require us to solve a series of systems of nonlinear equations (called penalized equations). In this paper, we first study the relationships among the solutions of penalized equations under appropriate conditions. Additionally, since these penalized equations are neither smooth nor convex, some existing algorithms, such as Newton method, cannot be applied directly to solve them. We shall apply the nonlinear Jacobian method to solve penalized equations and verify that the iteration sequence generated by the method converges monotonically to the solution of the penalized equation. Some numerical results confirm the theoretical results and the efficiency of the proposed algorithm.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Power penalty methods for the parabolic linear complementarity problem arising from American options pricing problems have attracted much attention; see, for example [1, 7, 12, 14–16]. In these methods, the parabolic linear complementarity problem is approximated by a nonlinear parabolic partial differential equation (which is called a penalized equation) via adding a penalty term. It is shown that the solution of the penalized equation converges to that of the parabolic linear complementarity problem with an arbitrary order and this allows us to achieve the required accuracy of the solution with a small penalty parameter [10, 12]. After the penalized equation is discretized in space and time, at each timestep it yields the following nonlinear discrete penalized equation: Find a vector \(W^k_\lambda \in \mathbb {R}^n\) such that
where \(A=(a_{ij})\in \mathbb {R}^{n\times n}\) is an M-matrix, \(f=(f_1,\ldots ,f_n)^T\in \mathbb {R}^n, g=(g_1,\ldots ,g_n)^T\in \mathbb {R}^n, \lambda >1\) is the penalty parameter, \(k\ge 1\) is the power exponent and \([u]_+=\max \{0,u\}\). The solution of the penalized equation \(W^k_\lambda \) is closely related to \(\lambda \) and \(k\). A matrix \(B\) is called an M-matrix if it has non-positive off-diagonals and \(B^{-1}\ge 0\) (i.e., all entries of \(B^{-1}\) are nonnegative) [8]. The penalized equation (1.1) can be regarded as the penalized formulation of the following discrete linear complementarity problem
In this paper, we shall study the relationships among the solutions of penalized equations and the relationships between the solution of the penalized equation and that of the discrete linear complementarity problem.
In order to price American options numerically, large-scale systems of nonlinear algebraic equations need to be solved at each discrete time point. When \(k=1\) in (1.1), the penalized equation is a semismooth equation and semismooth Newton method can be used to solve it [2, 4, 9]. When \(k>1\), the penalized equation (1.1) is a nonsmooth equation and classical or semismooth Newton method can not be applied directly. In [12], Wang et al. proposed a smoothing technique for the penalty term and with this technique, a damped Newton method was applied to solving the penalized equation. Compared with the cases where \(k=1\), it requires much more computational time to solve the penalized equation by the damped Newton method [14].
In this paper, we shall apply the nonlinear Jacobian iteration (see, e.g., [8]) to the penalized equation (1.1). A nonlinear SOR iteration was used to solve the discrete penalized equation arising from obstacle problems in [3] when \(k=1\). The nonlinear SOR iteration typically converges faster than the nonlinear Jacobi iteration by using the most recently available approximations of the elements of the iteration vector. However, the method is inherently sequential and it does not possess natural parallelism. Under proper conditions, we shall verify that the sequence generated by the nonlinear Jacobian iteration monotonically converges to the solution of the penalized equation.
The paper is organized as follows. In Sect. 2, we shall discuss convergence properties of solutions of penalized equations. In Sect. 3, we shall apply the nonlinear Jacobi iteration for solving the penalized equation and study the convergence property of the iteration. In Sect. 4, we shall discuss the solution of the subproblem. Finally, in Sect. 5, numerical experiments are presented to confirm the efficiency of the proposed method and to verify the theoretical results.
2 Convergence properties of solutions of penalized equations
We call \(W\) a lower-solution for the penalized equation (1.1) if
we also call \(W\) a lower-solution for the discrete linear complementarity problem (1.2) if
For two index subsets \(I, J\subset \{1,\ldots ,n\}\), let \(A_{IJ}\) be the submatrix of \(A\) with rows in \(I\) and columns in \(J\) and \(W_I\) be the subvector of \(W\) with entries indexes in \(I\). A matrix \(B\) (or a vector \(W\)) is nonnegative, denoted by \(B\ge 0\) (or \(W\ge 0\)), if all its entries are nonnegative. With these notions, we present some important relationships among the solutions of penalized equations and the discrete linear complementarity problem in the following lemmas.
Lemma 2.1
Let \(\lambda >1, k>0\) and \(W_\lambda ^k\) be the solution of the penalized equation (1.1). If \(\widetilde{W}\) is a lower-solution of the penalized equation (1.1), then \(\widetilde{W}\le W_{\lambda }^k\).
Proof
Since \(\widetilde{W}\) is a lower-solution of the penalized equation (1.1), it holds
This means that
Assume that there exist two disjoint nonempty index subsets \(I_1\) and \(I_2\) of \(\{1,\ldots ,n\}\) such that
From (2.1) it follows
Noting that \(A\) is an M-matrix, we obtain that \(A_{I_2I_1}\le 0\) and \(A_{I_2I_2}\) is also an M-matrix. From (2.2) we deduce
This is a contradiction to (2.2). The contradiction implies that \(I_2=\emptyset \). Therefore,
The proof is complete.\(\square \)
Lemma 2.2
Let \(1<\lambda _1<\lambda _2\) and \(W_{\lambda _1}^k\) be the solution of the penalized equation
Then \(W_{\lambda _1}^k\) is a lower-solution of the penalized equation
Moreover, \(W_{\lambda _1}^k\le W_{\lambda _2}^k\), where \(W_{\lambda _2}^k\) is the solution of (2.3).
Proof
Since \(1<\lambda _1<\lambda _2, [g-W_{\lambda _1}^k]_+\ge 0\) and \(W_{\lambda _1}^k\) is the solution of the penalized equation \( AW -\lambda _1[g-W]^{1/k}_+=f\), it must hold
This means that \(W_{\lambda _1}^k\) is a lower-solution of the penalized equation (2.3). Since \(W_{\lambda _2}^k\) is the solution of (2.3), by Lemma 2.1 we get \(W_{\lambda _1}^k\le W_{\lambda _2}^k\). The proof is complete.\(\square \)
Lemma 2.3
Let \(0<k_1<k_2\) and \(W_\lambda ^{k_1}\) be the solution of the penalized equation
Assume that for each \(i\), it holds \((W_\lambda ^{k_1})_i\ge g_i-1\). Then \(W_\lambda ^{k_1}\) is a lower-solution of the penalized equation
Moreover, \(W_\lambda ^{k_1}\le W_\lambda ^{k_2}\), where \(W_\lambda ^{k_2}\) is the solution of (2.4).
Proof
Since \(0<k_1<k_2, 0\le [g_i-(W_\lambda ^{k_1})_i]_+\le 1\) and \(W_\lambda ^{k_1}\) is the solution of the penalized equation \( AW -\lambda [g-W]^{1/k_1}_+=f\), it must hold
This means that \(W_\lambda ^{k_1}\) is a lower-solution of the penalized equation (2.4). Noting that \(W_\lambda ^{k_2}\) is the solution of (2.4), from Lemma 2.1 we deduce \(W_\lambda ^{k_1}\le W_\lambda ^{k_2}\).\(\square \)
Lemma 2.4
Let \(\lambda >1\) and \(k>0\). Assume that \(W_\lambda ^k\) is the solution of the penalized equation (1.1). Then \(W_\lambda ^k\) is a lower-solution of the discrete linear complementarity problem (1.2). Moreover, \(W_\lambda ^k\le W^*\), where \(W^*\) denotes the solution of (1.2).
Proof
Define two index sets \(I_1\) and \(I_2\) as follows
For each \(i\in I_1\), by (1.1) we have \(( AW _\lambda ^k)_i=f_i\), which means that \(\min \{(W_\lambda ^k-g)_i,\ ( AW _\lambda ^k-f)_i\}=0\); and for each \(i\in I_2\), we have \((W_\lambda ^k-g)_i<0\), which means that \(\min \{(W_\lambda ^k-g)_i,\ ( AW _\lambda ^k-f)_i\}<0\). This implies that for each \(i\), it holds
And hence \(W_\lambda ^k\) is a lower-solution of the discrete linear complementarity problem (1.2).
Noting that \(W^*\ge g\), by (2.5) we have \(W^*_i>\left( W_\lambda ^k\right) _i\) for each \(i\in I_2\). On the other hand, we have
This means that
Since \(A_{I_1I_1}\) is an M-matrix, we have \((W^*-W_\lambda ^k)_{I_1}\ge 0\). And hence \(W_\lambda ^k\le W^*\).\(\square \)
The following theorem shows that the sequence of solutions \(\{W_{\lambda _m}^k\}\) of the penalized equations is monotonically increasing and convergent to the solution of the discrete linear complementarity problem (1.2) as \(\lambda _m\rightarrow +\infty \).
Theorem 2.1
Let \(k\) be fixed and \(\{\lambda _m\}\) be a monotonically increasing sequence tending to positive infinity as \(m\rightarrow +\infty \). Assume that \(W_{\lambda _m}^k\) is the solution of the penalized equation
Then the sequence \(\{W_{\lambda _m}^k\}\) is monotonically increasing and convergent to the solution of the discrete linear complementarity problem (1.2). Moreover, when \(A\) is positive definite, there exists a constant \(C>0\), independent of \(n, W^*, W_{\lambda _m}^k\) and \(\lambda _m\), such that
where \(W^*\) is the solution of the discrete linear complementarity problem (1.2).
Proof
It follows from Lemmas 2.2 and 2.4 that
This implies that there exists some \(W^{**}\) such that \(\lim \nolimits _{m\rightarrow +\infty }W_{\lambda _m}^k=W^{**}\). Note that
Letting \(m\rightarrow +\infty \) in (2.7), we get \( AW ^{**}-f\ge 0\). On the other hand, we have
Letting \(m\rightarrow +\infty \) in (2.8), we get \([g-W^{**}]_+=0\). This means that \(W^{**}\ge g\). And hence it holds
By (2.7), we get
Letting \(m\rightarrow +\infty \), we obtain \(( AW ^{**}-f)^T(W^{**}-g)\le 0.\) This together with (2.9) implies that
Thus the above discussion shows that \(W^{**}\) is a solution of (1.2). Noting that \(A\) is an M-matrix, the linear complementarity problem (1.2) has a unique solution. Therefore, \(W^{**}=W^*\). The upper error bound (2.6) comes from Theorem 2.1 of [13] directly. The proof is complete.\(\square \)
Remark 2.1
-
(i)
The constant \(C\) in (2.6) is independent of the dimensionality \(n\) [13]. The upper error bound (2.6) can be further improved by the fact that all norms in a finite dimensional space are equivalent (see [5, 13] for details), i.e.,
$$\begin{aligned} \left\| W^*-W_{\lambda _m}^k\right\| _2\le \frac{C}{(\lambda _m)^k}. \end{aligned}$$But the constant \(C\) here is dependent of the dimensionality \(n\).
-
(ii)
In general, the sequence of solutions \(\{W_\lambda ^{k_m}\}\) of the penalized equations may not converge to the solution of the discrete linear complementarity problem if we let \(\lambda \) be fixed and \(k_m\rightarrow +\infty \). The following example has shown this. Consider the linear complementarity problem
$$\begin{aligned} W^*-3\ge 0,\quad 2W^*-2\ge 0\quad \mathrm{and}\quad (W^*-3)(2W^*-2)=0. \end{aligned}$$and the corresponding penalized equation
$$\begin{aligned} 2W_2^k-2\left[ 3-W_2^k\right] _+^{1/k}=2, \end{aligned}$$where \(\lambda =2\) and \(k>0\). By a simple calculation, we get \(W_2^k=2\) and \(W^*=3\). Therefore, the sequence of solutions \(W_2^{k_m}\) cannot converges to the solution as \(k_m\rightarrow +\infty \).
Theorem 2.2
Let \(A\) be positive definite, \(\lambda >\Vert Ag-f\Vert _1\) be fixed and \(\{k_m\}\) be a monotonically increasing sequence tending to positive infinity as \(m\rightarrow +\infty \). Assume that \(W_\lambda ^{k_m}\) is the solution of the penalized equation
If \(\lambda \) is chosen such that for each \(i, g_i\le (W_\lambda ^{k_1})_i+1\), then the sequence \(\{W_\lambda ^{k_m}\}\) is monotonically increasing and convergent to the solution of the discrete linear complementarity problem (1.2).
Proof
From Lemmas 2.3 and 2.4 it follows that the sequence \(\{W_\lambda ^{k_m}\}\) is monotonically increasing and bounded. As a result, there exists some \(W^{**}\) such that \(\lim \nolimits _{m\rightarrow +\infty }W_\lambda ^{k_m}=W^{**}\). Note that
Letting \(m\rightarrow +\infty \) in (2.10), we get \( AW ^{**}-f\ge 0\). On the other hand, we have
where \(p=1+1/{k_m}\). We decompose \(g-W_\lambda ^{k_m}\) into \(g-W_\lambda ^{k_m}=[u_1^T,u_2^T]^T\) with \(u_1>0\) and \(u_2\le 0\). Then the matrix \(A\) can be decomposed into \(A=\left( \begin{array}{ll}A_{11}&{}A_{12}\\ A_{21}&{}A_{22}\end{array}\right) \) and (2.11) can be rewritten as
Since \(u_1^TA_{11}u_1\ge 0\) and \(u_1^TA_{12}u_2\ge 0\), it holds
Using the Hȯlder’s inequality, we get
Note that \(\lambda >\Vert Ag-f\Vert _1\). Letting \(m\rightarrow +\infty \) in (2.12), we get \([g-W^{**}]_+=0\). This means that \(W^{**}\ge g\). Since \( AW ^{**}-f\ge 0\), it holds
By (2.10), we get
Letting \(m\rightarrow +\infty \), we obtain \(( AW ^{**}-f)^T(W^{**}-g)\le 0\), which together with (2.13) implies that
Thus, \(W^{**}\) is a solution of (1.2). Noting that the linear complementarity problem (1.2) has a unique solution, we get \(W^{**}=W^*\). The proof is complete.\(\square \)
Theorem 2.3
Let \(W^*\) be the solution of the discrete linear complementarity problem (1.2). Assume that \(W_\lambda ^k\) is the solution of the penalized equation (1.1). Define the following two index sets
Then for sufficiently large \(\lambda \), we have
and
Here, constants \(C_1, C_2\) and \(C_3\) only depend on \(k\) and \(\Vert AW ^*-f\Vert _\infty \).
Proof
For any \(i\in I_1\), by (1.1) we obtain
Noting that \(W_\lambda ^k-W^*\rightarrow 0\) as \(\lambda \rightarrow \infty \), we have \(|(A(W_\lambda ^k-W^*))_i|<\max \{( AW ^*-f)_i,1\}\) for \(\lambda \) sufficiently large. This implies that
Let \(C_1=2^k[\max \{\Vert AW ^*-f\Vert _\infty ,1\}]^k\). Then it holds
For any \(i\in I_2\), by (1.1) we get \(( AW ^k_\lambda -f)_i=0\). Since \(W^*_i\ge (W^k_\lambda )_i\), it holds \(W^*_i>g_i\) and hence \(( AW ^*-f)_i=0\). Therefore, we have
which implies that
Note that \(W_\lambda ^k\rightarrow W^*\) as \(\lambda \rightarrow \infty \). For sufficiently large \(\lambda \), we have \((W^k_\lambda )_i>g_i\) as long as \(W_i^*>g_i\). This means that \(W^*_{I_1}=g_{I_1}\) and hence
Let \(C_2=C_1\Vert (A_{I_2I_2})^{-1}A_{I_2I_1}\Vert _\infty \). Then it holds
Let \(C_3=C_1\max \{\Vert (A_{I_2I_2})^{-1}A_{I_2I_1}\Vert _\infty ,1\}\). From (2.14) and (2.15), it follows
The proof is complete.\(\square \)
Using Theorem 2.3, we get the following corollary.
Corollary 2.1
Let \(\lambda _1<\lambda _2\), and \(W_{\lambda _1}^k\) and \(W_{\lambda _2}^k\) be the corresponding solutions of the penalized equation (1.1). Then there exists some constant \(C\) such that
3 Nonlinear Jacobi iteration for solving the penalized equation
In this section, we shall apply the nonlinear Jacobi iteration for the solution of the discrete penalized equation (1.1) and analyze its convergence. For this purpose, we introduce some notations. We denote by \(D=\mathrm{diag}(a_{11},a_{22},\ldots ,a_{nn})\) the diagonal matrix composed of the diagonal elements of \(A\). Then, \(B\triangleq A-D\) denotes the matrix composed of the off-diagonal elements of \(A\). With these notations, we give the steps of the algorithm as follows.
Algorithm 3.1.
-
Step 1. Choose an initial point \(W^1\in \mathbb {R}^n\) and \(\epsilon >0\). Let \(\ell =1\).
-
Step 2. Compute \(W^{\ell +1}\) by solving the system of nonlinear equations
$$\begin{aligned} DW^{\ell +1}-\lambda \left[ g-W^{\ell +1}\right] _+^{1/k}=f- BW ^\ell . \end{aligned}$$(3.1) -
Step 3. Stop if \(\Vert W^{\ell +1}-W^\ell \Vert _\infty <\epsilon \). Otherwise, let \(\ell :=\ell +1\) and go to Step 2.
Remark 3.1
Let \(b^\ell = f- BW ^\ell \). Noting that \(D\) is a diagonal matrix, we can compute \(W^{\ell +1}\) in (3.1) by solving in parallel the following \(n\) one-dimensional nonlinear equations
The following lemma shows that the system (3.1) has a unique solution.
Lemma 3.1
For any \(\ell \ge 1\), the system of nonlinear equations (3.1) has a unique solution. Moreover,
Proof
It follows from Remark 3.1 that it only needs to prove the one-dimensional nonlinear equation (3.2) has a unique solution and \(W^{\ell +1}_i\ge b^\ell _i/a_{ii}\).
Consider the one-dimensional mapping \(h: R\rightarrow R\) defined by
Since the matrix \(A\) is an M-matrix, we have \(a_{ii}>0\). Thus, it is not difficult to verify that the mapping \(h\) is strictly monotone on \(R\) and
Hence the equation \(h(t)=0\) has a unique solution. This implies that (3.2) has a unique solution. On the other hand, by (3.2) we have
This means that (3.3) holds. The proof is complete.\(\square \)
We now give a convergence theorem of Algorithm 3.1.
Theorem 3.1
Let \(W^1\) be a lower-solution of the penalized equation (1.1). Then the sequence \(\{W^\ell \}\) generated by Algorithm 3.1 is monotonically increasing and convergent to the solution of the penalized equation (1.1).
Proof
The proof is divided into two parts.
We first verify that for each \(\ell =1,2,\ldots \), it holds
where \(W^k_\lambda \) is the solution of the penalized equation (1.1). Assume that \(W^\ell \) is a lower-solution. We shall prove that (3.4) holds for this \(W^\ell \). By Lemma 2.1 we only need to verify that \(W^{\ell +1}\) is also a lower-solution and \(W^\ell \le W^{\ell +1}\). Since \(W^\ell \) is a lower-solution, we have
Using the previous notations \(D, B\) and \(b^k\), it can be rewritten as
This implies that for each \(i=1,2,\ldots ,n\), it holds
Combining it with (3.2), we get
Since \(a_{ii}>0\), it is readily to see that \(W^\ell _i\le W^{\ell +1}_i\). Thus, \(W^\ell \le W^{\ell +1}\). On the other hand, we have
Noting that \(B\le 0\) and \(W^\ell \le W^{\ell +1}\), it follows from (3.1) that
This shows that \(W^{\ell +1}\) is also a lower-solution of the penalized equation (1.1). Since \(W^1\) is a lower-solution of the penalized equation (1.1), by the induction hypothesis we obtain (3.4) for each \(\ell \).
We now verify that the sequence \(\{W^\ell \}\) is convergent to the solution of the penalized equation (1.1). By (3.4), there exists some \(\bar{W}_\lambda ^k\) such that \(\lim \nolimits _{\ell \rightarrow +\infty }W^\ell =\bar{W}_\lambda ^k\). Letting \(\ell \rightarrow +\infty \) in (3.1), we get
and hence
This means that the sequence \(\{W^\ell \}\) is convergent to the solution of the penalized equation (1.1). The proof is complete.\(\square \)
4 The implementation of Algorithm 3.1
In Sect. 3, it has been established that the sequence generated by Algorithm 3.1 is monotonically convergent to the solution of the penalized equation if the initial iterate is a lower-solution. One question is how to choose the initial iterate, which is a lower-solution of (1.1). From Lemma 2.3, we know that the solution of the \(l_1\) penalized equation is a lower-solution of the corresponding \(l_k\) penalized equation. Thus, we can obtain the initial iterate via solving the \(l_1\) penalized equation. The other question is how to solve the one-dimensional nonlinear equation (3.2) effectively. To answer this question, we shall give a detailed discussion in the following.
We consider the following one-dimensional nonlinear equation: find \(U\in \mathbb {R}\) such that
where \(\alpha >0, \beta \in \mathbb {R}\) and \(\gamma \in \mathbb {R}\). Obviously, Eq. (4.1) is nonsmooth. Since \(\alpha >0\), by the discussion in Sect. 3 we know that (4.1) has a unique solution. In addition, \(U\ge \gamma /\alpha \) and the solution \(U\) satisfies the following equation
We shall study the analytical solution of Eq. (4.1) for the case where \(k=1\) and \(k=2\).
If \(\gamma /\alpha \ge \beta \), then \(U=\gamma /\alpha \). Otherwise, it holds
which yields that
Case I: \(k=1\). By a direct calculation, we get
Since \(\gamma /\alpha <\beta \), it holds \(U>\gamma /\alpha \) and it satisfies (4.1).
Case II: \(k=2\). Since \(\gamma /\alpha <\beta \), we have \(\varDelta \triangleq \lambda ^2+4\alpha ^2\beta -4\alpha \gamma >\lambda ^2\) and Eq. (4.2) has the following two solutions:
Noting that \(U\ge \gamma /\alpha \) and \(U_2<\gamma /\alpha \), we have \(U=U_1\) or equivalently
From the above discussion, we know that the one-dimensional nonlinear Eq. (3.2) has a unique analytical solution when \(k=1\) and \(k=2\). In fact, when \(k=3\) and \(k=4\), the equation (3.2) also has a unique analytical solution [6]. But its expression is more complicated and the details are omitted here. When \(k\ge 4\), we may apply the Newton-like iteration to solving the Eq. (4.1).
Note that the sequence \(\{W^\ell \}\) generated by Algorithm 3.1 is monotonically increasing and convergent to the solution of the penalized equation (1.1). It follows from Theorem 2.3 that we can choose the tolerance \(\epsilon =1/\lambda ^k\) in Step 3 of Algorithm 3.1.
5 Applications to American option pricing problems
In this section, we shall do some preliminary numerical experiments to verify the theoretical results and compare the performance of the proposed method with that of the method proposed in [12].
Consider an American put option with strike price \(K>0\) and maturity time \(T>0\). If the option is exercised when the underlying asset price is \(S\), the option holder receives the payoff
Let \(V(S,t)\) be the option value at time \(t\in [0, T]\) when the asset price is \(S\). Introducing a time-reverse transformation \(\tau =T-t, V(S,\tau )\triangleq V(S,T-t)\) satisfies the following form of the parabolic linear complementarity problem:
a.e. in \(\varOmega \triangleq (0,+\infty )\times [0, T)\), with the initial condition
and the differential operator
where \(\sigma \) is the volatility of the underlying asset and \(r>0\) is the risk free interest rate. For computational purposes, problem (5.1) will be posed on the localized domain
where \(S_\mathrm{max}\) denotes a sufficiently large number to ensure the accuracy of the solution, and we impose the following boundary conditions \(V(0,\tau )=\varPsi (0)\) and \(V(S_\mathrm{max},\tau )=\varPsi (S_\mathrm{max})\) for \(\tau \in [0, T]\). From [12], the penalized form of (5.1) is
where \(\lambda >0\) and \(k>0\).
5.1 Discretization
Following [1, 11, 12], we shall apply the finite volume discretization in space to (5.2). To the end, we let the interval \(I = (0, S_\mathrm{max})\) be divided into \(N\) subintervals,
with \(S_i=ih\) and \(h=S_\mathrm{max}/N\). We also let
These midpoints form a second partition of \((0, S_\mathrm{max})\) if we define \(S_{-1/2}=x_0\) and \(S_{N+1/2}=x_N\). Integrating both sides of (5.2) over \((S_{i-1/2}, S_{i+1/2})\), we get
for \(i=1,2,\ldots ,N-1\), where
Applying the midpoint quadrature rule to the first, third, and last terms, we obtain
where \(\varPsi _i=\varPsi (S_i), V_i(\tau )\) is the nodal approximation to \(V(S_i, \tau )\) to be determined, and \(\rho (V)\) is a flux associated with \(V\) defined by
To derive an approximation to the flux at the two endpoints \(S_{i-1/2}\) and \({S_{i+1/2}}\), let us consider the following two-point boundary-value problem
with
Denote \(\alpha \triangleq b/a=2(r/\sigma ^2-1)\). Solving the two-point boundary-value problem exactly, we obtain
which provides an approximation to the flux \(\rho (V)\) at \({S_{i+1/2}}\). Similarly, we can define an approximation of the flux at \(S_{i-1/2}\).
The above analysis does not apply to the approximation of the flux on \((0, x_1)\) since the two-point boundary-value problem is degenerate. To overcome this difficulty, let us reconsider it with an extra degree of freedom in the following form
with
where \(C\) is a constant to be determined. This local problem has the following solution [12]
which yields the following approximation to the flux \(\rho (V)\) in \(I_0\)
Substituting (5.4) and (5.5) into (5.3), we have
where
for \(i=2,3,\ldots ,N-1\). These form an \(N-1\) nonlinear ODE system for \(V(\tau )\triangleq (V_1(\tau ),\ldots , V_{N-1}(\tau ))^T\) with the homogeneous boundary condition \(V_0(\tau )=\varPsi (0)\) and \(V_N(\tau )=\varPsi (S_\mathrm{max})\). Using \(e_{i,j}\), we define an \((N-1)\times (N-1)\) tridiagonal matrix \(E\triangleq (e_{i,j})\) and an \((N-1)\)-dimensional vector
Then (5.6) can be rewritten as
where \(\varPsi =(\varPsi _1,\ldots ,\varPsi _{N-1})^T\). Equation (5.7) is a first order linear ODE system. To discretize it, we let \(\tau _i, i=0, 1, \ldots , M\), be a set of partition points in \([0, T]\) satisfying \(0=\tau _0\le \tau _1\le \ldots \le \tau _M=T\). Then, the application of the implicit timestepping scheme to (5.7) yields
where \(\varDelta \tau _m=\tau _{m+1}-\tau _m>0, V^{m+1}=V(\tau _{m+1})\) and \(f^{m+1}=V(\tau _{m+1})\). Letting
where \(I\) is an identity matrix, (5.8) reduces to (1.1). By [12], the matrix \(A\) is an M-matrix.
5.2 Numerical experiments
In this subsection, we set the parameters as follows:
We first verify the monotone convergence of the power penalty method and compare the orders of convergence rate for \(k=1\) and \(k=2\). When \(k=1\), we call the method the \(l_1\) penalty method. When \(k=2\), we call the method the \(l_{1/2}\) penalty method. In this experiment, we choose the grid step \(h=0.3125\). The corresponding node number is \(3201\), that is, \(N=3200\). To determine numerically the computational orders of convergence rate of the penalty method with respect to \(\lambda \), we choose a sequence of \(\lambda _p\) with \(\lambda _p=2\lambda _{p-1}\) (\(p=2, 3, 4, 5, 6\)) and some given \(\lambda _1>1\). Table 1 (resp. Table 2) lists the value of American put options at \(S=K, t=0\) or \(\tau =T\) [denoted by \(V_p(100,0)\)] and \(\Vert [g-V^n]_+\Vert _\infty \) at \(t=0\) [denoted by \(\mathrm{norm}_p([g-V]_+\))] for each \(p\) and \(k=1\) (resp. \(k=2\)). We denote \(\varDelta _p V=V_p(100,0)-V_{p-1}(100,0)\), and
From Tables 1 and 2, we see that the computed rates of convergence is about 1 for the \(l_1\) penalty method and about 2 for the \(l_{1/2}\) penalty method. That is,
Moreover, the \(l_{1/2}\) penalty method generates a more accurate approximation to the true solution than the \(l_1\) penalty method does, which is well matched with the theoretical results obtained in Sect. 2. From Tables 1 and 2, we also see that the value of American option at \(S=K, t=0\) is monotonically increasing as the \(\lambda \) is increasing, which shows the monotone convergence of the methods.
In the second experiment, we compare the computational performance of smoothing Newton method proposed in [12] and Algorithm 3.1 when applied to solve the \(l_{1/2}\) penalized equation. Note that the solution of the \(l_1\) penalized equation with the same penalty parameter \(\lambda \) is a lower-solution of the \(l_{1/2}\) penalized equation. We choose it as an initial iterate for both algorithms. We consider the \(l_{1/2}\) penalized equation with \(N=200, 400, 800, 1600\). We set the penalty parameter \(\lambda =4\) for these penalized equations. Table 3 lists the average number of iterations (denoted by Iter) of two algorithms per timestep and the total CPU time (denoted by CPU). Besides these, the value of American put options at \(S=K, t=0\) [denoted by \(V(100,0)\)] is listed in Table 3.
Table 3 shows that the value of American put options at \(S=K, t=0\) solved by both algorithms is almost the same. But the smoothing Newton method requires much more CPU time than Algorithm 3.1 does. Therefore, the proposed method is much more efficient than the smoothing Newton method.
References
Angermann, L., Wang, S.: Convergence of a fitted finite volume method for the penalized Black–Scholes equation governing European and American Option pricing. Numer. Math. 106, 1–40 (2007)
Chen, X., Nashed, Z., Qi, L.: Smoothing methods and semismooth methods for nondifferentiable operator equations. SIAM J. Numer. Anal. 38, 1200–1216 (2000)
Deckelnick, K., Siebert, K.: \(W^{1,\infty }\)-convergence of the discrete free boundary for obstacle problems. IMA J. Num. Anal. 20, 481–498 (2000)
Forsyth, P.A., Vetzal, K.R.: Quadratic convergence for valuing American options using a penalty method. SIAM J. Sci. Comput. 23, 2095–2122 (2002)
Huang, C.C., Wang, S.: A power penalty approach to a nonlinear complementarity problem. Oper. Res. Lett. 38, 72–76 (2010)
Kurosh, A.G.: Higher Algebra. MIR Publishers, Moscow (1972)
Li, W., Wang, S.: Pricing American options under proportional transaction costs using a penalty approach and a finite difference scheme. J. Ind. Manag. Optim. 9, 365–389 (2013)
Ortega, J.M., Rheinboldt, W.C.: Iterative Solution of Nonlinear Equations in Several Variables. Academic Press, San Diego (1970)
Qi, L., Sun, J.: A nonsmooth version of Newton’s method. Math. Program. 58, 353–367 (1993)
Rubinov, A., Yang, X.Q.: Lagrange-Type Functions in Constrained Non-convex Optimization. Kluwer Academic, Boston (2003)
Wang, S.: A novel fitted finite volume method for the Black–Scholes equation governing option pricing. IMA J. Numer. Anal. 24, 699–720 (2004)
Wang, S., Yang, X.Q., Teo, K.L.: Power penalty method for a linear complementarity problem arising from American option valuation. J. Optim. Theory Appl. 129, 227–254 (2006)
Wang, S., Yang, X.Q.: A power penalty method for linear complementarity problems. Oper. Res. Lett. 36, 211–217 (2008)
Zhang, K., Yang, X.Q., Wang, S., Teo, K.L.: Numerical performance of penalty method for American option pricing. Optim. Methods Softw. 25, 737–752 (2010)
Zhang, K., Teo, K.L.: Convergence analysis of power penalty method for American bond option pricing. J. Glob. Optim. 56, 1313–1323 (2013)
Zhang, K., Teo, K.L.: A penalty-based method from reconstructing smooth local volatility surface from American options. J. Ind. Manag. Optim. 11, 631–644 (2015)
Acknowledgments
The authors would like to thank the anonymous referees for their valuable suggestions and comments.
Author information
Authors and Affiliations
Corresponding author
Additional information
The first author was supported by the National Nature Science Foundation of P. R. China (11201197 and 11126147), the Nature Science Foundation of Jiangxi (20132BAB211011), the Foundation of Department of Education Jiangxi Province (GJJ13204). The third author was supported by the Research Grants Council of Hong Kong (PolyU 5295/12E).
Rights and permissions
About this article
Cite this article
Sun, Z., Liu, Z. & Yang, X. On power penalty methods for linear complementarity problems arising from American option pricing. J Glob Optim 63, 165–180 (2015). https://doi.org/10.1007/s10898-015-0291-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10898-015-0291-6
Keywords
- American option pricing
- Linear complementarity problem
- Penalized equations
- Iterative method
- Monotone convergence