Abstract
In this note, we focus on the spectral analysis of large matrices coming from isogeometric approximations based on B-splines of the eigenvalue problem
where u(0) and u(1) are given. When considering the collocation case, global distribution results for the eigenvalues are available in the literature, despite the nonsymmetry of the related matrices. Here we complement such results by providing precise estimates for the extremal eigenvalues and hence for the spectral conditioning of the resulting matrices. In the Galerkin setting, the matrices are symmetric and positive definite and a more complete analysis has been conducted in the literature. In the latter case we furnish a further procedure that gives a highly accurate estimate of all the eigenvalues, starting from the knowledge of the spectral distribution symbol. The techniques involve dyadic decomposition arguments, tools from the theory of generalized locally Toeplitz sequences, and basic extrapolation methods.
Access provided by Autonomous University of Puebla. Download chapter PDF
Similar content being viewed by others
1 Introduction
In this note we consider the approximation of one-dimensional elliptic eigenvalue problems by using an isogeometric either Galerkin or collocation technique with B-splines [5]. We are interested in the eigenvalues of the large matrices stemming from the considered approximation processes. In particular, we address the problem of estimating the extremal eigenvalues and of providing efficient numerical procedures for computing a good approximation of all the eigenvalues.
In this direction, it has recently been proven that the resulting sequence of matrices, indexed with respect to the matrix size, have a canonical distribution (see [15, 17] and references therein), by using the theory of Generalized Locally Toeplitz (GLT) sequences [16].
We recall that every GLT sequence has an associated function, called the symbol, and that the uniform sampling of the symbol provides an asymptotic approximation of all the eigenvalues, if every matrix of the GLT sequence is Hermitian as it happens in our setting. However, in general the approximation is quite poor and recently some extrapolation techniques have been devised (see [9] and references therein). In the constant coefficient setting, when considering the Galerkin approach, the results presented in [10] are impressive in the sense that machine precision is obtained with very low computational cost, while for the variable coefficients further improvements are needed (see also [17]).
In the current note, when considering the Galerkin setting and variable coefficients, we propose a further numerical scheme for the computation of all the spectrum of large matrices by using the numerical computation of the eigenvalues for small matrices, underlying asymptotic expansions, and extrapolation methods as those in [9]. The numerical results are of the same quality as those produced by the best strategy in [17], especially when the problem coefficients are smooth and at least for low frequencies, which are those of highest interest for understanding the nature of the problem.
On the other hand, when dealing with the collocation approximation we obtain large nonsymmetric matrices. However, the GLT machinery can be employed and the symbol is real-valued and nonnegative [7], as in the symmetric positive definite Galerkin setting [15]. Hence we expect that most of the eigenvalues are real or with negligible imaginary part. Here we start the analysis of the collocation case, by describing a technique based on dyadic decompositions, for estimating the extreme eigenvalues and hence the asymptotic (spectral) conditioning of the involved matrix sequences. The analysis contained in Theorem 1 and Corollary 1 confirms that the conditioning grows at most as \(h^{-2}\), h being the fineness parameter, exactly as in the case of the matrices obtained with the Galerkin approximation (see e.g [14]).
Finally, it is not clear if the delicate asymptotic expansions observed in [10, 17] holds also in the collocation setting and indeed this issue will be the subject of future investigations.
2 Preliminaries
In the following we present the notation that we use. In particular we give the definition of eigenvalue distribution, that of rearrangement, and we briefly discuss the informal meaning behind these definitions.
A matrix-sequence is any sequence of the form \(\{\mathbf X_n\}_n\), where \(\mathbf X_n\) is a square matrix such that \(\mathrm{size}(\mathbf X_n)\rightarrow \infty \) as \(n\rightarrow \infty \). Let \(\mu _d\) be the Lebesgue measure in \(\mathbb R^d\) and let \(C_c(\mathbb C)\) be the space of continuous complex-valued functions with bounded support defined on \(\mathbb C\). If \(\mathbf X\) is an \(m\times m\) matrix, the eigenvalues of \(\mathbf X\) are denoted by \(\lambda _1(\mathbf X),\ldots ,\lambda _m(\mathbf X)\). If \(\mathbf g:D\subset \mathbb R^d\rightarrow \mathbb C^{s\times s}\) is an \(s\times s\) matrix-valued function, we say that \(\mathbf g\) is measurable if its \(s^2\) components \(g_{ij}:D\rightarrow \mathbb C\), \(i,j=1,\ldots ,s\), are measurable.
Definition 1
Let \(\{\mathbf X_n\}_n\) be a matrix-sequence, let \(N_n:=\mathrm{size}(\mathbf X_n)\), and let \(\mathbf g:D\subset \mathbb R^d\rightarrow \mathbb C^{s\times s}\) be a measurable \(s\times s\) matrix-valued function defined on a set D with \(0<\mu _d(D)<\infty \). We say that \(\{\mathbf X_n\}_n\) has an (asymptotic) eigenvalue distribution described by \(\mathbf g\), and we write \(\{\mathbf X_n\}_n\sim _\lambda \mathbf g\), if
where \(\lambda _i(\mathbf g(y_1,\ldots ,y_d))\), \(i=1,\ldots ,s\), are the eigenvalues of the \(s\times s\) matrix \(\mathbf g(y_1,\ldots ,y_d)\).
The informal meaning behind the eigenvalue distribution \(\{\mathbf X_n\}_n\sim _\lambda \mathbf g\) is the following: for all sufficiently large n, the eigenvalues of \(\mathbf X_n\) can be subdivided into s different subsets of approximately the same cardinality; and the eigenvalues belonging to the ith subset (except possibly for \(o(N_n)\) outliers) are approximated by the samples of the ith eigenvalue function \(\lambda _i(\mathbf g(y_1,\ldots ,y_d))\) over a uniform grid in D (the domain of \(\mathbf g\)). For example, if \(d=1,\ N_n=ns\) and \(D=[a,b]\), then the eigenvalues of \(\mathbf X_n\) are approximately equal to
for n large enough. Likewise, if \(d=2,\ N_n=n^2s\) and \(D=[a_1,b_1]\times [a_2,b_2]\), then the eigenvalues of \(\mathbf X_n\) are approximately equal to
for n large enough.
Remark 1
(rearrangement) Let \(\mathbf g:D\subseteq \mathbb R^d\rightarrow \mathbb C^{s\times s}\) and suppose that D is a rectangle in \(\mathbb R^d\), say \(D:=[a_1,b_1]\times \cdots \times [a_d,b_d]\). We also assume that the eigenvalues \(\lambda _1(\mathbf g(y_1,\ldots ,y_d)),\ldots ,\lambda _s(\mathbf g(y_1,\ldots ,y_d))\) are real for all \((y_1,\ldots ,y_d)\in D\). For each positive integer r, let \(\mathcal G_r\) be the uniform grid in D given by
Compute the samples of the eigenvalue functions
at the points \((y_1,\ldots ,y_d)\in \mathcal G_r\), sort them in increasing order and put them in a vector \((q_1,q_2,\ldots ,q_{sr^d})\). Let \(\eta _r:[0,1]\rightarrow \mathbb R\) be the piecewise linear non-decreasing function that interpolates the samples \((q_0:=q_1,q_1,q_2,\ldots ,q_{sr^d})\) over the nodes \((0,\frac{1}{sr^d},\frac{2}{sr^d},\ldots ,1)\), i.e.,
Under certain (normally satisfied) conditions on \(\mathbf g\), the function \(\eta _r\) converges (a.e.) as \(r\rightarrow \infty \) to a non-decreasing function \(\eta :[0,1]\rightarrow \mathbb R\), which is referred to as the rearranged version of \(\mathbf g\). What is important about \(\eta \) is that
Therefore, if we have \(\{\mathbf X_n\}_n\sim _\lambda \mathbf g\), then we also have \(\{\mathbf X_n\}_n\sim _\lambda \eta \).
3 Isogeometric Galerkin Discretization of Variable-Coefficient Eigenvalue Problems
Consider the following one-dimensional variable-coefficient eigenvalue problem:
where \(\varOmega \) is an open interval in \(\mathbb R\) and \(a,b\in L^1(\varOmega )\) are such that \(a,b>0\) a.e. on \(\varOmega \). The corresponding weak formulation reads as follows: find eigenvalues \(\lambda _j\in \mathbb R^+\) and eigenfunctions \(u_j\in H^1_0(\varOmega )\), for \(j=1,2,\ldots ,\infty \), such that, for all \(v\in H^1_0(\varOmega )\),
where
In the isogeometric Galerkin method, we assume that the physical domain \(\varOmega \) is described by a global geometry map \(G:[0,1]\rightarrow \overline{\varOmega }\), which is invertible and satisfies \(G(\partial ([0,1]))=\partial \overline{\varOmega }\). We fix a set of basis functions \(\{\varphi _1,\ldots ,\varphi _{N_n}\}\) defined on the reference (parametric) domain [0, 1] and vanishing on the boundary \(\partial ([0,1])\). We consider the basis functions
which are defined on the physical domain \(\varOmega \), and we define the approximation space \(\mathscr {W}_n:=\mathrm{span}(\psi _1,\ldots ,\psi _{N_n})\subset H^1_0(\varOmega )\). Finally, we find approximations to the exact eigenpairs \((\lambda _j,u_j)\), \(j=1,2,\ldots ,\infty \), by solving the following Galerkin problem: find \(\lambda _{j,n}\in \mathbb R^+\) and \(u_{j,n}\in \mathscr {W}_n\), for \(j=1,\ldots ,N_n\), such that, for all \(v_n\in \mathscr {W}_n\),
Assuming that both the exact and numerical eigenvalues are arranged in non-decreasing order, the pair \((\lambda _{j,n},u_{j,n})\) is taken as an approximation of the pair \((\lambda _j,u_j)\) for all \(j=1,\ldots ,N_n\). The numbers \(\lambda _{j,n}/\lambda _j-1,\ j=1,\ldots ,N_n\), are referred to as the (relative) eigenvalue errors.
In view of the canonical identification of each function \(v_{n}\in \mathscr {W}_n\) with its coefficient vector with respect to the basis \(\{\varphi _1,\ldots ,\varphi _{N_n}\}\), solving the Galerkin problem (2) is equivalent to solving the generalized eigenvalue problem
where \(\mathbf u_{j,n}\) is the coefficient vector of \(u_{j,n}\) with respect to \(\{\varphi _1,\ldots ,\varphi _{N_n}\}\) and
The matrices \(\mathbf K_n(a,G)\) and \(\mathbf M_n(b,G)\) are, respectively, the stiffness and mass matrices. Due to our assumption that \(a,b>0\) a.e. on \(\varOmega \), both \(\mathbf K_n(a,G)\) and \(\mathbf M_n(b,G)\) are always symmetric positive definite, regardless of the chosen basis functions \(\varphi _1,\ldots ,\varphi _{N_n}\) and the map G. Moreover, it is clear from (3) that the numerical eigenvalues \(\lambda _{j,n}\), \(j=1,\ldots ,N_n\), are just the eigenvalues of the matrix
Therefore, if \(a, b\in L^1(\varOmega )\) and the basis functions \(\varphi _1,\ldots ,\varphi _{N_n}\) are chosen as the B-splines of degree p and global smoothness \(C^k\) defined on the interval [0, 1] partitioned into n equal subintervals (see [17, Sect. 2.1]) and denoted by
then
From the *-algebra structure of GLT sequences, as proven in [16, 17], we know that
where, according to Remark 1, \(\kappa _{[p,k]}\), \(\xi _{[p,k]}\), \(\zeta _{[p,k]}\) are the rearranged versions of
respectively, with
where the blocks \(\mathbf K_{[p,k]}^{[\ell ]}\) and \(\mathbf M_{[p,k]}^{[\ell ]}\), of size \(p-k\), are defined in [17]. The analytical predictions of the eigenvalue errors are obtained through the following sampling procedure:
where \(m:=\min (n(p-k),n(p-k)+k-1)\).
However, such a procedure is not completely satisfactory especially in low frequencies (see [17]), when variable coefficients are considered. More precisely, we recall that the smallest eigenvalues are those related to slowly oscillating modes, that is the related eigenvectors belong to the subspace generated by low frequencies. Since the eigenvalues of the continuous operator appear in the denominator in formula (16), the relative error could be higher even if the absolute approximation is of high precision.
We address this issue in the next section.
For notational completeness, we stress that in the following the geometrical map G is always the identity and hence the matrices we consider are those denoted as \(\mathbf L_n(a,b):= \mathbf L_n(a,b,I)\) with \(\mathbf L_n(a,b,G)\) as in (6); furthermore, the regularity will be maximal that is \(k=p-1\) and hence \(e_p(\theta )\), \(f_p(\theta )\), \(h_p(\theta )\) will denote \(\mathbf e_{[p,p-1]}(\theta ), \mathbf f_{[p,p-1]}(\theta ), \mathbf h_{[p,p-1]}(\theta )\), respectively.
4 Global Distribution Results and Extrapolation
As discussed in [17], the mismatch between the analytical predictions and the eigenvalue errors essentially occurs only for small eigenvalues and a way to significantly reduce it has already been illustrated in that paper. Drawing inspiration from [1, 9, 10, 12, 13], we here describe—in the case of an isogeometric Galerkin discretization based on B-splines of the eigenvalue problem (1)—an alternative interpolation–extrapolation procedure to considerably improve the analytical predictions for small eigenvalues. We also illustrate the performance of this procedure by two examples.
-
1.
We assume that there exists a function \(c_p:[0,\pi ]\rightarrow \mathbb R\), depending only on p and the coefficients a, b of the considered eigenvalue problem (1), such that, when using an isogeometric p-degree \(C^{p-1}\) B-spline discretization, the following property holds: independently of the parameter n, all the eigenvalues of the matrix \(\frac{1}{n^2}\mathbf L_n(a,b)\), sorted in increasing order, satisfy
$$\begin{aligned} \frac{\lambda _{j,n}}{n^2}\approx c_p(\theta _{j,n})e_p(\theta _{j,n}),\qquad j=1,\ldots ,\min (n+p-2,n), \end{aligned}$$(17)where
$$\begin{aligned} e_p= \frac{f_p(\theta )}{h_p(\theta )}&=\frac{(2-2\cos \theta )h_{p-1}(\theta )}{h_p(\theta )}\\&=\frac{(2-2\cos \theta )\bigl (\phi _{[2p-1]}(p)+2\sum _{\ell =1}^{p-1}\phi _{[2p-1]}(p-\ell )\cos (\ell \theta )\bigr )}{\phi _{[2p+1]}(p+1)+2\sum _{\ell =1}^p\phi _{[2p+1]}(p+1-\ell )\cos (\ell \theta )} \end{aligned}$$is defined in [17], and
$$ \theta _{j,n}=\frac{j\pi }{n},\qquad j=1,\ldots ,n. $$ -
2.
Compute the eigenvalues of the matrix \(\frac{1}{n_1^{2 {L}}}\mathbf {L}_{n_1}(a,b)\) corresponding to a small value of \(n_1\), i.e.,
$$ \frac{\lambda _{j_1,n_1}}{n_1^2},\qquad j_1=1,\ldots ,n_1+p-2. $$Since \(n_1\) is small, this computation can be efficiently performed by any standard eigensolver (e.g., the Matlab eig function).
-
3.
Under the assumption in item 1,
$$\begin{aligned} c_p(\theta _{j_1,n_1})\approx \frac{\lambda _{j_1,n_1}}{n_1^2e_p(\theta _{j_1,n_1})},\qquad j_1=1,\ldots ,\min (n_1+p-2,n_1). \end{aligned}$$This means that we have an approximation of the unknown function \(c_p\) over the coarse uniform grid consisting of the points \(\theta _{j_1,n_1}\).
-
4.
Interpolate the data
$$ \left( \theta _{j_1,n_1},\frac{\lambda _{j_1,n_1}}{n_1^2e_p(\theta _{j_1,n_1})}\right) ,\qquad j_1=1,\ldots ,\min (n_1+p-2,n_1), $$by using, e.g., the Matlab interp1 function with the ‘spline’ option, so as to obtain an approximation of \(c_p\) over the whole interval \([0,\pi ]\). We call this approximation \(\tilde{c}_p\). Note that, by construction,
$$ \tilde{c}_p(\theta _{j_1,n_1})=\frac{\lambda _{j_1,n_1}}{n_1^2e_p(\theta _{j_1,n_1})},\qquad j_1=1,\ldots ,\min (n_1+p-2,n_1). $$ -
5.
Given a large n, compute approximations of the numerical eigenvalues \(\lambda _{j,n}\) by replacing \(c_p\) with \(\tilde{c}_p\) in (17):
$$ \lambda _{j,n}\approx n^2\tilde{c}_p(\theta _{j,n})e_p(\theta _{j,n}),\qquad j=1,\ldots ,\min (n+p-2,n). $$ -
6.
Compute analytical predictions for the eigenvalue errors as follows:
$$ \frac{\lambda _{j,n}}{\lambda _j}-1\approx \frac{n^2\tilde{c}_p(\theta _{j,n})e_p(\theta _{j,n})}{\lambda _j}-1,\qquad j=1,\ldots ,\min (n+p-2,n). $$
The next two examples show that the interpolation–extrapolation algorithm defined by items 1–6 may lead to analytical predictions of the eigenvalue errors which perform better (for small eigenvalues) than the analytical predictions proposed in (16).
Example 1
Let \(p=1\), \(n=200\), \(a(x)=2+0.5\,x\) and \(b(x)=1\). Let \(n'=1500\gg n\), \(p'=5\), and take the first \(n-1\) eigenvalues of \(\mathbf L_{n'}(a,b)\), namely \(\lambda _{1,n'},\ldots ,\lambda _{n-1,n'}\), as approximations of the unknown exact eigenvalues \(\lambda _1,\ldots ,\lambda _{n-1}\). In Table 1 we report the five smallest eigenvalues and approximated eigenvalues for \(p=1\). As is clear the new approximation, \(n^2\tilde{c}_1(\theta _{j,n})e_1(\theta _{j,n})\), performs better than \(n^2\zeta _{r}\left( \frac{j}{n}\right) \) to approximate \(\lambda _{j,n}\), where the function \(\zeta _{r}\) introduced in [17, Sect. 3.1] is defined in the following way. Sample \(\frac{a(x)}{b(x)}e_1(\theta )\) at the grid points \((x,\theta )\in \mathcal {G}_r\), for a chosen r, where
The samples are ordered in increasing order in a vector \((z_1,z_2,\ldots ,z_{r^2})\). Let \(\zeta _r:[0,1]\rightarrow \mathbb {R}\) be a piecewise linear non-decreasing function that interpolate the samples \((z_0:=z_1,z_2,\ldots ,z_{r^2})\) over the nodes \((0,\frac{1}{r^2},\frac{2}{r^2},\ldots ,1)\).
In Fig. 1 we present the (approximate) analytical predictions \(n^2\zeta _r(\frac{j}{n})/\lambda _{j,n'}-1\), with \(r=\) 10000, together with the (approximate) eigenvalue errors \(\lambda _{j,n}/\lambda _{j,n'}-1\). We note the mismatch for the smallest eigenvalues. In Fig. 2 we plot the (approximate) analytical predictions \(n^2\tilde{c}_1(\theta _{j,n})e_1(\theta _{j,n})/\lambda _{j,n'}-1\), obtained from the above interpolation–extrapolation algorithm for \(n_1=10\), and the (approximate) eigenvalue errors \(\lambda _{j,n}/\lambda _{j,n'}-1\) versus \(j/(n-1)\), for \(j=1,\ldots ,n-1\). We clearly see that in Fig. 2 the slight mismatch for small eigenvalues observed Fig. 1 has been reduced.
Example 2
Let \(p=2,3,4,5\), \(n=200\), \(a(x)=2.1\cdot 10^9+1.05\cdot 10^9\,x\) and \(b(x)=8000\). The approximation parameters \(n'\) and \(n_1\) have been chosen as \(n'=1500\) and \(n_1=10\), respectively, and the eigenvalues \(\lambda _{j,n'}\) have been chosen correspondingly, in the sense that \(\lambda _{j,n'}\) is computed as follows: we compute the eigenvalues \(\lambda _{j,n'}\) for a matrix \(\mathbf {L}_{n'}(a,b)\) of order \(n'=1500\) and B-spline degree \(p'=5\). Then we assume that the first \(n+p-2\) (for \(n=200\)) of these eigenvalues, that is \(j=1,\ldots ,n+p-2\), are accurate representations of the true eigenvalues \(\lambda _j\); see [17] for a motivation and a discussion on this matter.
In Table 2 we present the five smallest eigenvalues and approximated eigenvalues for \(p=2,3,4,5\). As is clear the new approximation, \(n^2\tilde{c}_p(\theta _{j,n})e_p(\theta _{j,n})\), performs better than \(n^2\zeta _{r,[p,p-1]}\left( \frac{j}{n}\right) \) to approximate \(\lambda _{j,n}\). Figure 3 shows the comparison between the (approximate) analytical predictions
and the (approximate) eigenvalue errors
whereas Fig. 4 shows the comparison between the (approximate) analytical predictions
with the (approximate) eigenvalue errors. We see from Fig. 4 that the mismatch for small eigenvalues observed in Fig. 3 has been lowered.
5 Dyadic Decomposition Argument and Extreme Eigenvalues
While the distribution results are available both in the Galerkin [15] and in the collocation setting [7], the use of extrapolation methods, as those described in the previous section and in [10, 11], has been developed only in the Galerkin setting. The reason is the inherent nonsymmetry of the collocation matrices. However, this issue has to be investigated in the future, even if few preliminary experiments seem quite promising. In this section, we start by analyzing the extreme eigenvalues of the (nonsymmetric) stiffness matrices in the collocation setting. We use a dyadic decomposition argument already employed for symmetric structures in several contexts (see [3, 20] and references therein).
Consider the one-dimensional Poisson problem
where \(\mathrm{f}\in C([0,1])\). Suppose we approximate (18) by using the isogeometric collocation method based on uniform B-splines (see [7, Sect. 4] for the details on this method). Then, the resulting discretization matrix is
where \(\xi _{i,[p]},\ i=2,\ldots ,n+p-1,\) are the Greville abscissae defined in [7, Eq. (4.6)], while \(N_{j,[p]},\ j=2,\ldots ,n+p-1,\) are the usual B-spline basis functions of maximal regularity \(k=p-1\). In other words, with reference to (7), we have
We denote by \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[p]})\) the real part of \(\tilde{\mathbf {K}}_n^{[p]}\) (the real part of a square complex matrix X [4] is by definition \((X+X^{\mathrm {H}})/2\), \(X^{\mathrm {H}}\) being the transpose conjugate). In Theorem 1 we prove for \(p=3\) the following result:
where
and
Theorem 1
There exists a constant \(c_3>0\) such that
Proof
For \(n=2,\ldots ,5\), a direct verification shows that \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[3]})\) is positive definite. Therefore, the theorem is proved if there exists a constant \(c>0\) such that
For \(n\ge 6\) and \(c>0\), the matrix \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[3]})-c\,\tau _{n+1}(2-2\cos \theta )\) is explicitly given by
and can be decomposed as follows:
In the above decomposition, r, s, and t are arbitrary real numbers, whereas each term of the summation is a \((n+1)\times (n+1)\) matrix whose only nonzero entries are shown in the associated box and are contained in the principal submatrix corresponding to the rows from the superscript to the subscript. For instance,
If each term of the above summation is a nonnegative definite matrix, then \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[3]})-c\,\tau _{n+1}(2-2\cos \theta )\) is nonnegative definite as well. The following conditions ensure that each term of the summation is nonnegative definite:
These conditions are satisfied, for instance, with
Hence, (21) holds with \(c=58/59\). \(\square \)
We verified that the dyadic decomposition argument used in the proof of Theorem 1 can also be used to prove (20) for \(p=2,4\). Although this argument becomes quite difficult to apply for \(p\ge 5\), we have reason to believe that a careful application of it could prove (20) for any given \(p\ge 2\).
The above theorem and the relations in (20) have interesting consequences on the conditioning measured with respect to the induced Euclidean norm as reported in Corollary 1, whose proof requires a technical but interesting in itself lemma.
Lemma 1
Given X square matrix, assuming \(\mathfrak {R}(X)\) positive definite, we have
Proof
As a first step define the following matrix \(\hat{X}\) of order 2n
By taking into account the standard singular value decomposition, it is easy to see (and it is well-known) that the singular values \(\sigma _j(X)\) of X, \(j=1,\ldots ,n\), coincide with the n largest eigenvalues of \(\hat{X}\), since the eigenvalues of the Hermitian matrix \(\hat{X}\) are \(\pm \sigma _j(X)\), \(j=1,\ldots ,n\). Therefore, we infer that
with \(\lambda _{1}(\hat{X})\le \lambda _{2}(\hat{X})\le \cdots \le \lambda _{2n}(\hat{X})\). From the minimax characterization (see [4]) of the \((n+1)\)-th eigenvalue of \(\hat{X}\), we obtain
where the maximum is taken over all subspaces U with codimension n in \({\mathbb {C}}^{2n}\) or, equivalently, over all subspaces U of dimension n. Let \(U_*\) be the n-dimensional subspace of \({\mathbb {C}}^{2n}\) made of all vectors y of the form
If \(y\in U_*\) is partitioned into blocks as above, it holds
and since \(U_*\) is a particular subspace of dimension n, from (22) the thesis follows. \(\square \)
Corollary 1
Assume that (20) is satisfied for a given p. Then
and
with \(C_p\) a positive constant depending only on p and \(\mu (X)=\Vert X\Vert \Vert X^{-1}\Vert \) being the conditioning of an invertible matrix X measured with respect to the induced Euclidean norm \(\Vert \cdot \Vert \).
Proof
First of all we observe that both \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[p]})\) and \(\tilde{\mathbf {K}}_n^{[p]}\) are banded matrices with coefficients not depending on the parameter n. Hence we have that their induced Euclidean norm is asymptotic to a a constant not depending on n that is
We first prove (24). By (20) we have
with \(c_p>0\) and hence \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[p]})\) is positive definite and its minimal eigenvalue is bounded from below by
\(4\sin ^2\left( \frac{\pi }{2(n+p-1)}\right) \) being the minimal eigenvalue of \(\tau _{n+p-2}(2-2\cos \theta )\) (see [19] and references therein). On the other hand, \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[p]})\) contains as a principal submatrix a Toeplitz matrix of size proportional to n and generated by \(f_p\) being nonnegative and having a unique zero of order two (see [7, 14]). Hence such a Toeplitz matrix has the minimal eigenvalue \(l_n\) asymptotic to \(n^{-2}\) (see again [19] and references therein) and by the Cauchy interlacing theorem the minimal eigenvalue of \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[p]})\) is bounded from above by \(l_n\sim n^{-2}\). Therefore
But \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[p]})\) is positive definite and hence
from which, using (26), we deduce
We now prove (25). Given the known fact \(\Vert X^{-1}\Vert =[\sigma _{\min }(X)]^{-1}\), statement (25) is a consequence of (27), of (20), of Lemma 1, and of the fact that the minimal eigenvalue of \(\tau _{n+p-2}(2-2\cos \theta )\) is
\(\square \)
Furthermore, we have reasons to conjecture that the spectral conditioning of the collocation stiffness matrices grows asymptotically as \(n^2\), exactly as in the Galerkin setting [14], since \(\tilde{\mathbf {K}}_n^{[p]}\) contains as a principal submatrix a Toeplitz matrix of size proportional to n and generated by \(f_p\) having a unique zero of order two (as \(\mathfrak {R}(\tilde{\mathbf {K}}_n^{[p]})\), see the proof of Corollary 1).
We finally stress that (20) is also important in a multigrid context for proving the optimality of the related two-grid and multigrid techniques (see [2, 6, 8, 19]), by following the theory reported by Ruge and Stüben in [18].
6 Conclusions
In this note, we have considered the spectral analysis of large matrices coming from the isogeometric approximations based on B-splines of the eigenvalue problem
where u(0) and u(1) are given. In the collocation setting, we complemented global eigenvalue distribution results, available in the literature [7], with precise estimates for the extremal eigenvalues and hence for the spectral conditioning of the resulting matrices. In the Galerkin setting, we have designed an efficient matrix-less procedure (see [9]) that gives a highly accurate estimate of the all the eigenvalues, starting from the knowledge of the spectral GLT distribution symbol. Possible extensions include a more systematic treatment of the collocation case, both via the use of dyadic decomposition arguments and via the use of proper matrix-less extrapolation techniques. This last item is completely new and represents a real challenge, thus it will be the subject of a future investigation.
References
Ahmad, F., Al-Aidarous, E.S., Abdullah Alrehaili, D., Ekström, S.-E., Furci, I., Serra-Capizzano, S.: Are the eigenvalues of preconditioned banded symmetric Toeplitz matrices known in almost closed form? Numer. Algorithms 78, 867–893 (2018)
Aricò, A., Donatelli, M., Serra-Capizzano, S.: V-cycle optimal convergence for certain (multilevel) structured linear systems. SIAM J. Matrix Anal. Appl. 26, 186–214 (2004)
Beckermann, B., Serra-Capizzano, S.: On the asymptotic spectrum of finite element matrix sequences. SIAM J. Numer. Anal. 45, 746–769 (2007)
Bhatia, R.: Matrix Analysis. Springer, New York (1997)
Cottrell, J.A., Hughes, T.J.R., Bazilevs, Y.: Isogeometric Analysis: Toward Integration of CAD and FEA. Wiley, Chichester (2009)
Donatelli, M., Garoni, C., Manni, C., Serra-Capizzano, S., Speleers, H.: Two-grid optimality for Galerkin linear systems based on B-splines. Comput. Vis. Sci. 17, 119–133 (2015)
Donatelli, M., Garoni, C., Manni, C., Serra-Capizzano, S., Speleers, H.: Spectral analysis and spectral symbol of matrices in isogeometric collocation methods. Math. Comput. 85, 1639–1680 (2016)
Donatelli, M., Garoni, C., Manni, C., Serra-Capizzano, S., Speleers, H.: Symbol-based multigrid methods for Galerkin B-spline isogeometric analysis. SIAM J. Numer. Anal. 55, 31–62 (2017)
Ekström, S.-E.: Matrix-less methods for computing eigenvalues of large structured matrices. Ph.D. Thesis, Uppsala University (2018)
Ekström, S.-E., Furci, I., Garoni, C., Manni, C., Serra-Capizzano, S., Speleers, H.: Are the eigenvalues of the B-spline isogeometric analysis approximation of \(-{\varDelta } u=\lambda u\) known in almost closed form? Numer. Linear Algebr. Appl. 25, e2151 (2018)
Ekström, S.-E., Furci, I., Serra-Capizzano, S.: Exact formulae and matrix-less eigensolvers for block banded symmetric Toeplitz matrices. BIT Numer. Math. 58, 937–968 (2018)
Ekström, S.-E., Garoni, C.: A matrix-less and parallel interpolation-extrapolation algorithm for computing the eigenvalues of preconditioned banded symmetric Toeplitz matrices. Numer. Algorithms 80(3), 819–848 (2019)
Ekström, S.-E., Garoni, C., Serra-Capizzano, S.: Are the eigenvalues of banded symmetric Toeplitz matrices known in almost closed form? Exp. Math. 27, 478–487 (2018)
Garoni, C., Manni, C., Pelosi, F., Serra-Capizzano, S., Speleers, H.: On the spectrum of stiffness matrices arising from isogeometric analysis. Numer. Math. 127, 751–799 (2014)
Garoni, C., Manni, C., Serra-Capizzano, S., Sesana, D., Speleers, H.: Spectral analysis and spectral symbol of matrices in isogeometric Galerkin methods. Math. Comput. 86, 1343–1373 (2017)
Garoni, C., Serra-Capizzano, S.: Generalized Locally Toeplitz Sequences: Theory and Applications. Springer Monographs in Mathematics, vol. 1. Springer, Cham (2017)
Garoni, C., Speleers, H., Ekström, S.-E., Hughes, T.J.R., Reali, A., Serra-Capizzano, S.: Symbol-based analysis of finite element and isogeometric B-spline discretizations of eigenvalue problems: Exposition and review. Arch. Comput. Methods Eng. (In press) https://doi.org/10.1007/s11831-018-9295-y
Ruge, J.W., Stüben, K.: Algebraic multigrid. In: McCormick, S. (ed.) Multigrid Methods, pp. 73–130. SIAM Publications, Philadelphia (1987)
Serra-Capizzano, S.: Convergence analysis of two-grid methods for elliptic Toeplitz and PDEs matrix-sequences. Numer. Math. 92, 433–465 (2002)
Serra-Capizzano, S., Tablino-Possio, C.: Spectral and structural analysis of high precision finite difference matrices for elliptic operators. Linear Algebra Appl. 293, 85–131 (1999)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Ekström, SE., Serra-Capizzano, S. (2019). Eigenvalue Isogeometric Approximations Based on B-Splines: Tools and Results. In: Giannelli, C., Speleers, H. (eds) Advanced Methods for Geometric Modeling and Numerical Simulation. Springer INdAM Series, vol 35. Springer, Cham. https://doi.org/10.1007/978-3-030-27331-6_4
Download citation
DOI: https://doi.org/10.1007/978-3-030-27331-6_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-27330-9
Online ISBN: 978-3-030-27331-6
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)