Abstract
Using sampling experiments, we found that, when there are fewer groups than variables, between-groups PCA (bgPCA) may suggest surprisingly distinct differences among groups for data in which none exist. While apparently not noticed before, the reasons for this problem are easy to understand. A bgPCA captures the g − 1 dimensions of variation among the g group means, but only a fraction of the \(\sum {n_{i} } - g\) dimensions of within-group variation (\(n_{i}\) are the sample sizes), when the number of variables, p, is greater than g − 1. This introduces a distortion in the appearance of the bgPCA plots because the within-group variation will be underrepresented, unless the variables are sufficiently correlated so that the total variation can be accounted for with just g − 1 dimensions. The effect is most obvious when sample sizes are small relative to the number of variables, because smaller samples spread out less, but the distortion is present even for large samples. Strong covariance among variables largely reduces the magnitude of the problem, because it effectively reduces the dimensionality of the data and thus enables a larger proportion of the within-group variation to be accounted for within the g − 1-dimensional space of a bgPCA. The distortion will still be relevant though its strength will vary from case to case depending on the structure of the data (p, g, covariances etc.). These are important problems for a method mainly designed for the analysis of variation among groups when there are very large numbers of variables and relatively small samples. In such cases, users are likely to conclude that the groups they are comparing are much more distinct than they really are. Having many variables but just small sample sizes is a common problem in fields ranging from morphometrics (as in our examples) to molecular analyses.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
As a general trend, modern science tends to generate a very large number of variables (p) from samples that can vary widely in size (n) and often includes few individuals relative to the number of variables. Indeed, the ‘Omics’ revolution, brought forward by the rapid advancement of informatics and molecular biology, offers some of the best examples of this trend. For instance, microarray analyses may include hundreds of genetic markers from a relatively small number of individuals (Culhane et al. 2002 is an example). However, statistically analyzing such high dimensional data with relatively small sample sizes (p/n ratios) is an important and challenging problem.
A variety of methods for dimensionality reduction are available in the statistical literature (Izenman 2008). Among these, principal component analysis (PCA) is still probably the most popular in biology. A PCA is a rigid rotation of the multidimensional space of all the variables followed by a projection of the data onto relatively few orthogonal axes that together account for as much of the overall variance as possible, though there is no reason for the axes themselves to be especially meaningful biologically. When p ≥ n, a PCA can only extract at most n − 1 uncorrelated dimensions that, together, contain all the information about the variances and covariances of the original p variables (all p dimensions can be extracted when p < n). Often, there are dominant directions of variance so that a relatively small number of PCs may account for most of the variation. The first (higher order) PCs capture the major aspects of covariation in the sample and the later PCs the smaller ones. Bookstein (2017) first brought attention to the Marchenko-Pastur theorem that shows that large p/n ratios cause an exaggeration of the sizes of the eigenvalues for the first PCs relative to those of the last PCs, thus giving a misleading impression of the relative importance of the patterns that they seem to suggest. The initial motivation for the present paper was to investigate whether large p/n ratios might cause problems for the relatively new and increasingly popular type of PCA, between-group PCA (bgPCA). In this method a PCA is performed on the covariance matrix based on the g sample means (rather than on the original data matrix) followed by the projection of the original n samples onto these bgPC axes. Plots of these axes are then used to illustrate the distances between sample means and allow a user to judge the distinctiveness of the groups.
Phenotypic variation is complex and, although the number and choice of morphometric descriptors should be determined by the specific study hypothesis (Oxnard and O’Higgins 2011), morphometric studies are often exploratory, tending to employ large numbers of variables, which make this discipline typically highly multivariate (Blackith and Reyment 1971). This is intrinsically true for landmark coordinate-based GM (geometric morphometrics), because each additional landmark or semilandmark adds two variables to a 2D study or three to a 3D study. While the p/n ratios are very variable (Table 1), datasets used in GM studies often have many more measurements than specimens. This is particularly common in, but not exclusive to, anthropology, the discipline in which semilandmark methods for the analysis of curves and surfaces were developed and are widely employed to study human evolution (Bookstein 1997; Gunz and Mitteroecker 2013; Slice 2005). Semilandmarks are typically closely spaced sets of arbitrary points used to ‘discretize’ anatomical features, such as curves and surfaces, that are devoid of clearly corresponding landmark points; therefore, they can greatly increase the number of variables in a study. Indeed, a propensity for morphometrics to employ large numbers of variables has become especially evident in the last decade, thanks to new, cheaper and faster instruments for the acquisition and analysis of 3D images. For instance, almost 60% of about 1000 entries, retrieved at the end of 2018 in Publish or Perish (https://harzing.com/resources/publish-or-perish) using google scholar to search “geometric morphometrics AND semilandmarks”, were papers published since 2013.
Description of the bgPCA Method
An important topic in biology is the description and interpretation of group differences in multivariate spaces Various approaches have been suggested to summarize among group variation in scatterplots (ordination methods) and to classify individuals in groups. Yet, today’s most commonly multivariate technique for separating groups is still multi-group linear discriminant analysis (DA), also known as canonical variates analysis (CVA), originally proposed by Fisher (1936) and Mahalanobis (1936). However, a limit for using DA/CVA in a study is that, for statistical reliability, it requires sample sizes greatly exceeding the count of variables in the analysis (Mitteroecker and Bookstein 2011), and indeed it is not even computationally defined if \(p > n - g\). In these instances, a between-group PCA (bgPCA) has been suggested as an interesting potential alternative to explore group structure. To our knowledge, this method was originally proposed by Yendle and MacFie (1989) who called it “discriminant principal components analysis” (DPCA), though it does not involve a standardization by the within-group variation as in DA and CVA. Another early paper is Culhane et al. (2002), who applied it to the analysis of high-dimensional microarray data. While bgPCA has similarities with discriminant functions, but also, as discussed by Boulesteix (2005), has relationships to partial least-squares dimension reduction methods. Compared to DA/CVA, bgPCA is just a PCA and does not involve standardizing the variables based on the variation within groups (Seetah et al. 2012). Also, as with DA/CVA, bgPCA has been used for classification, and thus for predicting group affiliation based on bgPCs, an aim which should be achieved with a cross-validation, as exemplified by leave-one-out jack-knife used in Culhane et al. (2002) and Seetah et al. (2012). However, in contrast to a DA/CVA (Kovarovic et al. 2011; Mitteroecker and Bookstein 2011), a bgPCA does not require \(p \le n - g\), which is why it has been claimed that “in… between-group PCA there is NO restriction on the number of variables” (https://www.mail-archive.com/morphmet@morphometrics.org/msg05221.html).
The bgPCA procedure is used to reduce the dimensionality of multivariate data to just those dimensions necessary to account for the differences among the g group means. Each sample is based on \(n_{i}\) individuals for a total sample size of \(n = \sum {n_{i} }\) or \(n = gn_{i}\) in the case of equal sample sizes, as will be assumed here for simplicity. A bgPCA is performed by projecting the original n × p data matrix, X, onto the matrix, E, of the normalized eigenvector of the among-group SSCP matrix \({\mathbf{A}} = \sum\limits_{i}^{g} {n_{i} } \left( {{\bar{\mathbf{x}}}_{i} - {\mathbf{\bar{\bar{x}}}}} \right)^{t} \left( {{\bar{\mathbf{x}}}_{i} - {\mathbf{\bar{\bar{x}}}}} \right)\), where \({\bar{\mathbf{x}}}_{i}\) is the row vector for the mean of the ith group and \({\mathbf{\bar{\bar{x}}}}\) is the grand mean vector. The A matrix is at most of rank g − 1 because it is a PCA of just the matrix of g means so only the first g-1 eigenvalues can be greater than zero and thus only the first g − 1 columns of E need to be retained. The n × (g − 1) transformed data matrix is then \({\mathbf{X^{\prime}}} = {\mathbf{XE}}\). Based on these, the transformed within-group and among-group SSCP matrices are \({\mathbf{W^{\prime}}} = {\mathbf{E}}^{t} {\mathbf{WE}}\) and \({\mathbf{A^{\prime}}} = {\mathbf{E}}^{t} {\mathbf{AE}} = {\varvec{\Lambda}}\), the diagonal matrix of the first g − 1 eigenvalues of A (note: the superscript “t” indicates matrix transpose; also, while the equation for A given above weights the mean for each group by its sample size, that may not be appropriate for many applications, see Bookstein 2019, but it is used here for generality). Importantly, the number of bgPCs cannot be more than g − 1. Thus, with just two groups, there are only two group means, and one needs a single dimension to represent differences between two points; thus, when g = 2, there is only one bgPC. If there are three groups, the differences among the three corresponding means can be fully described by a plane passing through the three mean points, and thus by just two bgPCs. With g > 3 the rationale is the same and the number of bgPCs is g − 1, but the geometric representation is not as easy, because we cannot represent multivariate spaces with more than three dimensions in a single scatterplot and even a 3D scatterplot (as with g = 4) can be difficult to interpret (Mitteroecker et al. 2005).
Sampling Experiments
To investigate the effect of varying p/n ratios on bgPCA, sampling experiments were performed using both isotropic data (independent variables with equal means and variances called Model 1 below) and data constructed from an actual morphometric study but with no true differences among the group means (called Models 2–3 below). Figure 1 shows the result of bgPCAs using g = 3 groups with the same true means (i.e., no real group differences), a constant total sample size (n = 120), and an increasingly larger numbers of variables (p = 12, 120 or 360). On the left (Fig. 1a) are bgPCA plots for isotropic data (Model 1, below) for g = 3 groups of identical size (\(n_{i} = {n \mathord{\left/ {\vphantom {n 3}} \right. \kern-0pt} 3} = 40\)). On the right (Fig. 1b), the same \(n_{i}\), g and p are used as in Fig. 1a but based on correlated morphometric variables from real data, which have been randomly divided into three groups so that there are no real group differences. Convex hulls for each group are shown in order to identify group memberships for each sample. Rather than showing the groups superimposed as one might expect, because there are no true differences, Fig. 1 shows that bgPCA created an apparent clustering of the samples around their group means as first noticed by one of us (AC). The groups appear increasingly distinct from one another as the p/n ratio increases because larger numbers of variables are used. The effect is particularly evident for isotropic data and less pronounced but still present for correlated variables.
The sampling experiments, shown in Fig. 1, were based on two different models, one (Fig. 1a) being the same as model 1 (below) and the other (Fig. 1b) being similar to models 2–3 (below). In all instances, there are no true differences among the means of the groups and the groups have the same size. Thus, in more detail, the models used in the more extensive sampling experiments described below, were:
- Model 1:
-
A purely isotropic model with p independent random normally distributed variables, each with μ = 0 and σ = 1. This model was used for Figs. 1a, 2, and 4 below
- Models 2&3:
-
Random normally distributed variables with the same true covariance matrix as that of a real morphometric dataset, but with all means equal to zero:
- Model 2:
-
Procrustes shape coordinates from a sample of 45 adult yellow-bellied marmot (Marmota flaviventris) left hemimandibles. The original 2D configuration consists of 10 landmarks and 50 semilandmarks, with the semilandmarks slid in TPSRelw (Rohlf 2015) using the minimum Procrustes distance criterion. This data matrix was then used to compute the covariance matrix among the variables and its corresponding eigenvector matrix and eigenvalues. All eigenvectors that had positive eigenvalues were retained. These were then used as described below to generate random data matrices with the covariance matrix taken from the original dataset
- Model 3:
-
Procrustes shape coordinates from a sample of 171 adult male vervet monkey skulls, which are part of a larger published dataset (Cardini and Elton 2017). There were 86 3D skull landmarks (Cardini et al. 2007; Cardini and Elton 2017). As with Model 2, as described below, these were used to generate samples of random variables with the same true covariance matrix as in the original data
A sample, X, from a population with a given true covariance matrix of Σ was generated using the following relationship. \({\mathbf{X}} = {\mathbf{YE\varLambda }}^{{{1 \mathord{\left/ {\vphantom {1 2}} \right. \kern-0pt} 2}}}\), where Y is an n × p matrix of independent random normally distributed numbers with zero means and unit variances, E is a matrix of the p, p-dimensional normalized eigenvectors of Σ, and Λ is the p × p diagonal matrix of its eigenvalues. A difference between sampling experiments using the isotropic model (Model 1) and all others based on actual data (Models 2 and 3) is that the maximum number of eigenvectors that can be computed is limited to the number of variables in the original study because the method cannot construct more dimensions than are in the original data. For models 2 and 3, random samples of the rows (corresponding to the variables) of matrix E were used to generate variables. When the desired p was greater than the original number of variables, variables were obtained by sampling the rows of E with replacement.
In the sampling experiments that follow, the data were subjected to a bgPCA using code written by FJR in MATLAB and group separation was assessed by computing an index of overlap between pairs of samples. Let \(O_{ij}\) be the proportion of individuals in a group i that are closer to the mean of group j. When the dispersions in two groups i and j do not overlap, \(O_{ij}\) will be equal to 0 and will approach 0.5 for a pair of groups that overlap almost perfectly, because in that case a point is equally likely to be closest to either mean. The average, \(\bar{O}_{ij}\) for all pairs of samples in a particular analysis is used as the measure of overlap. Initially, the amount of overlap between convex hulls was considered, but this has some unsuitable properties (such as rapid decrease in the probability of overlap as the number of dimensions increases even without the bgPCA transformation).
What Happens When n or g are Changed Relative to p?
Figure 2 summarizes the results of sampling experiments using \(\bar{O}_{ij}\) as a measure of overlap and varying g, \(n_{i}\), and p. The figure uses \(n_{i}\) rather than n because the total size is not relevant for the computation of average overlap as they depend on the relationships among pairs of samples and not the number of samples (and thus not on the total sample size). The sampling experiments used g = 3 and 6 groups, sample sizes of \(n_{i}\) = 20 and 40, and a range of values for the number of dimensions, p. Figure 2 shows the expected outcome that overlap is larger when p is smaller, \(n_{i}\) larger, and when there are more groups. The effect of p is strongest for the isotropic model, but the effect is clear for all three models. The companion paper also demonstrates the effect of relaxing the assumption of equal sample sizes.
Mathematical Interpretation: Why the Apparent Separation of Groups as p Increases?
Because the \(\bar{O}_{ij}\) index seems difficult to work with analytically, an alternative index inspired by the partitioning of sums of squares in an anova or MANOVA was investigated for the simple null model (Model 1) used above, i.e., samples of independent normally distributed random variables from the same population. As an approximation, covariances among the variables are ignored (as they should be minimal for isotropic data) and the group differences described in terms of the traces (sums of the diagonal elements) of the usual within and among-groups sums of squares matrices, rather than the usual multivariate test statistics such as Wilks’ Lambda or Lawley–Hotelling U statistics, which require the computation of the matrix inversion and determinants of the sums of squares matrices.
The reader should carefully note that all expressions in Table 2 are based just on the g − 1-dimensional space of the bgPCA transformed data. Thus, the within-group sums of squares here only refers to that part of total within group sums of squares expected in the g − 1-dimensional subspace. This table is not intended for and should never be used for statistical testing (unlike that of a standard MANOVA, which would use the variation in the p-dimensional space of the original variables even if resampling procedures are used), and is specifically designed to produce an explanation for the apparent differences between groups such as shown Fig. 1a.
As above, let A represent the among-groups SSCP matrix based on all p variables and E its matrix of normalized eigenvectors. After projecting the data for all samples onto these vectors, one has a bgPCA transformed data matrix \({\mathbf{X^{\prime}}} = {\mathbf{XE}}\). At most, only the first g-1 columns of E and thus \({\mathbf{X^{\prime}}}\) are nonzero, so we will use only the first g-1 columns. Let \({\mathbf{A^{\prime}}}\) be the among-groups SSCP matrix based on this transformed data matrix. The sum of the eigenvalues of A and \({\mathbf{A^{\prime}}}\) are equal because all of the variation among g means is captured in a g-1-dimensional space. Similarly, one can define W as the within-groups SSCP matrix using the original p variables and \({\mathbf{W^{\prime}}}\) as the equivalent matrix using the projections of the data onto E. Note that its trace \(tr\left( {{\mathbf{W^{\prime}}}} \right)\) will, in general, be less than that of W because only within-group variation in the g − 1 dimensions in which the means differ is preserved by the projection onto the g − 1-dimensional bgPCA space. The W matrix has n-g degrees of freedom and thus would require \(\hbox{min} \left( {n - g,p} \right)\) dimensions to account for all the within-group variation.
Consider sampling experiments, such as described in the prior section for Model 1, where \(n_{i}\) specimens are in each sample (assuming equal sample size, so that \(n = gn_{i}\)) are drawn from the same p-dimensional multivariate normal distribution, that has a mean vector \({\varvec{\upmu}} = {\mathbf{0}}_{p}\) (a vector of p zeroes) and a covariance matrix of \({\varvec{\Sigma}} = {\mathbf{I}}_{p}\) (a p × p identity matrix). The true W matrix would then be \(\left( {n - g} \right){\mathbf{I}}_{p}\) with \(tr\left( {\mathbf{W}} \right) = p\left( {n - g} \right)\). The true among groups variance component matrix, \({\varvec{\Sigma}}_{A}\), is \({\mathbf{0}}_{p}\) because there are no true differences among the population means. However, due to sampling error the expected among-groups covariance matrix is \({\varvec{\Sigma}} + n_{i} {\varvec{\Sigma}}_{A}\). For the transformed data, the trace of the observed among-groups SSCP matrix is unchanged by the transformation because all of the variation among g means will be accounted for by the g − 1 eigenvectors. However, the trace of the expected within-groups SSCP matrix will be reduced by the fraction \({{\left( {g - 1} \right)} \mathord{\left/ {\vphantom {{\left( {g - 1} \right)} p}} \right. \kern-0pt} p}\) assuming the remaining n–g dimensions of within-group variation are just a random sample of the total variation (reasonable here because, as mentioned above, there are no actual differences). These relations are conveniently summarized in the format of a MANOVA table (Table 2), but just using the trace of each matrix divided by g − 1 as a summary of the relative amounts of within and among samples variation captured in the bgPCA space only.
Note that the \(F_{iso}\) ratio defined in Table 2 (ratio of traces of among to within group MS using only the g − 1 bgPCs) is analogous to an F-ratio and is a function of just p and g. The subscript “iso” is to remind the reader that it assumes isotropic data and is not the usual F employed for statistical testing (that, as mentioned, should not be done using the equations of Table 2). Likewise, the “iso” in the subscript of \(R_{iso}^{2}\) is to remind the reader that this is not the usual squared multiple correlation coefficient, because this statistic, as it is computed here using only the bgPCA variance, is only aimed at assessing the amount of apparent group separation. Thus, a value near zero would imply that groups account for little of the total variation and values near 1 imply that most of the variation is between groups rather than within groups. Figure 3 shows plots of \(R_{iso}^{2} = {p \mathord{\left/ {\vphantom {p {\left( {p + g\left( {n_{i} - 1} \right)} \right)}}} \right. \kern-0pt} {\left( {p + g\left( {n_{i} - 1} \right)} \right)}}\) as a function of \(n_{i}\) and p for g = 3 and 6 that illustrate how \(R_{iso}^{2}\) increases as a function of p (suggesting more distortion with more variables), but decreases as a function of \(n_{i}\) (indicating less separation of groups with larger samples). For a given p and \(n_{i}\), if g is smaller, and therefore also \(n = gn_{i}\) is smaller, the denominator in the \(R_{iso}^{2}\) formula is reduced and \(R_{iso}^{2}\) becomes larger, which is why the \(R_{iso}^{2}\) surfaces in Fig. 3 are higher for g = 3 than for g = 6. This is because adding more groups increases the dimensionality of the bgPCA space and thus should account for a larger proportion of the within-group variation.
The reader should note that larger \(R_{iso}^{2}\) implies more separation and thus less overlap as measured by \(\bar{O}_{ij}\). Figure 4 shows a scatterplot \(\bar{O}_{ij}\) as a function of \(R_{iso}^{2}\) using the data from Fig. 2. The slope of the relationship differs for data from the different models. The slope is less steep for the models with correlated variables. Within each dataset the scatter corresponds to the effects of different values of g and \(n_{i}\). The \(R_{iso}^{2}\) statistic is somewhat ad hoc, but Fig. 4 shows that it is a useful predictor of overlap for isotropic data.
The expressions in Table 2 are compared in Table 3 with the results from two sampling experiments. The example in the upper half is for the case where there are fewer variables but larger sample sizes in each group. The second for the case where the number of variables is larger and sample sizes are smaller. The values are averages over 10,000 replications and show the close agreement with the expected values (given in parentheses) computed using the formulas from Table 2.
The Effect of Covariation Among Variables
The isotropic Model 1, used in the previous section, is based on the unrealistic assumption that the variables are independent and have equal variances. Intuitively, one might expect that data with highly correlated variables might be less prone to overestimating of the degree of group separation, and indeed the sampling experiments presented in Figs. 1b, 2 and 4 do show less spurious separation for data with correlated variables (i.e., the models using vervet and marmot covariance matrices). If, as an extreme case, because of a strong correlation between variables, all of the variation in a dataset could be accounted for with just g − 1 dimensions, then all of the within-group variation would also be captured by the g − 1 among-groups dimensions of the bgPCA and no information would be lost. The \(R_{iso}^{2}\) statistic described above should then be close to 0 and \(\bar{O}_{ij}\) should measure the correct amount of overlap between groups, which should be close to 0.5 if there are no real groups).
In order to investigate the effect of covariation using sampling experiments, one must specify a model for the pattern and strengths of the correlations. The selection of a model can be simplified because one can rotate the data matrix to its principal axes, so that one need only consider models that differ in how the eigenvalues decrease as a function of their number. For independent variables they would decrease somewhat according to the Marchenko–Pastur formula (Bookstein 2017), but for highly correlated variables they would decrease more rapidly. A very simple model is that the logs of the eigenvalues, \(\ln \left( {\lambda_{i} } \right)\), decrease linearly as a function of the log of their number, that is, \(\ln \left( {\lambda_{i} } \right) = a - b\ln \left( i \right)\) or as \(\lambda_{i} = e^{ - bi}\), where a is a constant greater than 0 (ignored here) and b determines how rapidly the eigenvalues decrease. This approach also models the effect of unequal variances for the different variables. More realistic models with a factor structure could have been investigated, but this model seems sufficient to illustrate the effect of different proportions of the variance being accounted for by the first g − 1 dimensions. Figure 5a shows examples with b varied from 0 to 1. Larger values of b yield increasingly rapid declines of successive eigenvalues, which imply stronger correlations among variables.
Figure 5b shows the results of, sampling experiments with g = 3 groups of \(n_{i}\) = 20 observations each, with p ranging from 3 to 80, and each replicated 1000 times, for the b values used in Fig. 5a. The effect of increasing correlations among the variables was to reduce the size of the expected \(R_{iso}^{2}\) statistic implying a larger \(\bar{O}_{ij}\) and thus less spurious clustering of points around the means. Many morphometric datasets follow patterns like that shown for b equal to 1 or even larger values of b. For instance, the curve for the marmot mandible dataset (Model 2) would be even more extreme than the curve shown for b = 1 The curve for the vervet data (Model 3) is less extreme. Thus, it is not surprising that Fig. 1 shows that for data with highly correlated variables there will be much less spurious group separation than that found for the isotropic model (Model 1).
Discussion
The primary focus of the present paper is on the reasons for the apparent clustering of points around the means of arbitrary groups and predicting the magnitude of this distorted summary of group differences. In contrast, the companion paper, Bookstein (2019), examines the effect of large p/n ratios on the bgPCA method in relationship to the predictions of the Marchenko-Pastur theorem as described in Bookstein (2017), along with two other aspects of the problem: the role of variations in sample sizes of the groups, and the effect of correlations among the variables based on a variety of factor models. It also suggests ways of evaluating the impact of these effects when analyzing actual data sets. We use sampling experiments and examples from our own field, morphometrics, i.e. the quantitative study of biological forms (Blackith and Reyment 1971, Bookstein 1991, 2018). However, the issue and its implications are general and apply similarly to multivariate data used to compare groups in other fields such as genetics.
In our analyses we found that bgPCA ordinations may tend to exaggerate differences between groups relative to the amount of within-group variation. In extreme cases, with few groups, small samples and very many variables, bgPCA may consistently show perfect separation of the groups even when there are no true differences among group means. This is in part because the g − 1 dimensions of a bgPCA capture the entire amount of variation among the g group means, but only a fraction of the variation within each group when p > g − 1. Thus, most of the variance within groups is lost, when p is much larger than g − 1. With small samples, the groups may appear quite distinct, but any apparent group differences will largely be an artefact of very large sampling error (Cardini and Elton 2007; Cardini et al. 2015). This is because any inaccuracies in group mean estimates are completely captured by the bgPCs, as if they were true differences, and used to define the g − 1-dimensional space.
Not surprisingly, one can also see in Fig. 2 that, with the same p and g, larger samples overlap more than smaller samples. Indeed, whether there are true differences or not, the range of variation within a sample is expected to increase as its sample size increases and thus there is a greater chance of overlapping.
In summary, the distortion showing a consistent spurious degree of separation between groups is not a promising property for a method that was proposed to analyze data with large numbers of variables and small samples, but the picture is complex, because the gravity of the problem, as nicely exemplified by Fig. 4, varies sharply from case to case. Indeed, the severity of the distortion depends on both g and \(n_{i}\) relative to p, as well as on how strongly variables covary and whether true differences are indeed present (a case which we did not explore in our simulations). This is not unlike what Kovarovic et al. (2011) found in a study of discriminant analysis (DA). They remarked that (p. 3012): “increasing the number of predictors may increase … group separation in scatterplots of non-cross-validated DFAs, even if those predictors are random numbers which do not add any relevant information on group differences”. However, with bgPCA, this well-known problem of DA may be even more serious, because in bgPCA there is no theoretical limit to the number of variables that can be used to summarize groups and thus p can be much larger than n and g, as in many publications (Table 1).
Among the factors that might reduce the distortion, or even make it negligible, covariance is one of the most interesting, as it is expected in most biological datasets. The reason why covariance mitigates against the problem of bgPCA spurious group separation is that, with correlated variables, the number of independent dimensions is effectively reduced and, therefore, operationally, it is as if the \({p \mathord{\left/ {\vphantom {p {g\left( {n_{i} - 1} \right)}}} \right. \kern-0pt} {g\left( {n_{i} - 1} \right)}}\) ratio was smaller. The degree to which it is smaller depends on the strength of the covariances. Yet, the problem is clearly still there, as both separation and \(R_{iso}^{2}\) still increase with p. Thus, the main conclusion is the same: even with covariances, with a large \({p \mathord{\left/ {\vphantom {p {g\left( {n_{i} - 1} \right)}}} \right. \kern-0pt} {g\left( {n_{i} - 1} \right)}}\) ratio, not only might one see groups that appear overly separated, as in our sampling experiments, but also, if there are true groups, the differences will be inflated by a case-specific degree, which is difficult to predict a priori.
There are many reasons to expect strong covariances in studies using Procrustes-based GM. Some covariance is introduced by the fact that, for 2D data, the superimposition reduces the 2q-dimensional variation of the raw coordinates (with q being the number of landmarks) to the 2q–4 dimensions of shape space (Rohlf and Slice, 1990). In addition, covariation will depend on factors such as the number and distribution of the anatomical points. For example, landmarks that are very close together and closely spaced semilandmarks are expected to be highly correlated (Cardini 2018). Thus, the marmot data includes slid semilandmarks and 90% of the total variance in these data can be accounted for by just the first 10 PCs (out of the 44 possible because n = 45 and p = 120). By contrast, the vervet data requires 56 PCs (out of the 170 possible because n = 171 and p = 251) to account for the same percentage of total variance. Figure 2 shows that the curves for the marmot data are higher (more overlap and thus less false clustering) than the curves for the vervet data (less overlap and thus stronger false separation of the groups). Note that these results do not suggest that one should purposely add highly correlated variables to reduce the distortion expected in the results of a bgPCA. Adding perfectly correlated variables to an existing dataset will not change the effective dimensionality of a dataset and thus will not alter the degree of false clustering expected in the results of a bgPCA.
On the other hand, in datasets where strong correlations among variables are expected, such as is common in GM, where additional covariance is introduced by the Procrustes superimposition itself (Rohlf and Slice, 1990) and many semilandmarks are used (because physically close semilandmarks tend to covary strongly), one might hope to circumvent some of the issues raised in this paper by reducing the number of variables used in the bgPCA. Indeed, in GM studies, it is often the case that distance matrices among specimens assessed using a few landmarks are highly correlated with those derived from the full set of landmarks plus many semilandmarks (Skinner et al. 2009; Ferretti et al. 2013; Watanabe 2018; Galimberti et al. 2019). This can be assessed formally, for instance, through matrix correlations where testing whether full (all landmarks and semilandmarks) and reduced (a subset of the full configuration) data matrices are highly correlated. Thus, smaller ratios of \({p \mathord{\left/ {\vphantom {p {g\left( {n_{i} - 1} \right)}}} \right. \kern-0pt} {g\left( {n_{i} - 1} \right)}}\) can be achieved at the outset, simply by limiting the number of variables used in the study. If this is done, the resulting visualizations of shape differences among specimens will be less detailed, because fewer landmarks are used, but results of bgPCA will be less likely to be misleading.
It is important to bear in mind that scatterplots are not the only tool for assessing group differences. Results from a bgPCA should be complemented by tests of significance, as well as by cross-validated classifications of groups (e.g., Seetah et al. 2012). However, they must be performed using the full p-dimensional space (unlike the statistics in the ‘ad-hoc’ MANOVA Tables 2 and 3, using only bgPCs with the specific aim of assessing the magnitude of spurious group differences in the bgPCA sub-space). With small samples, and/or negligible group separation in the full data space, group differences using all p variables will be non-significant, thus alerting the user that any appearance of group separation in bgPCA scatterplots should be regarded with extreme suspicion. Also, as one of the main aims in the formulation of bgPCA by Culhane et al. 2002 was classification, the results should be checked by cross-validating bgPCAs in the full data space. Finding a cross-validated accuracy only negligibly different from that expected by chance should warn users about likely distortions in the scatterplots.
In conclusion, big datasets are increasingly common, but having very many variables does not ‘counterbalance’ the effect of small n; it could make it worse, as shown here and in Bookstein (2019). Thus, we show that in attempting to assess group distinctiveness using bgPCA there is a potential trap, in that spurious apparent groupings may emerge in scatterplots, especially when the subspace spanned by the g-1 bgPCs does not adequately reflect within group variation, as is increasingly likely to happen when p/n is large and g is small. The appearance of spurious groups in bgPCA offers a good reminder of how a large number of descriptors might bring problems as well as benefits, with the problems sometimes potentially outweighing the benefits. Indeed, as with other methods (Hair et al. 2009; Bookstein 2017), bgPCA provides another example of the potential perils of high dimensional data, and of the possible misuse of techniques and misinterpretation of findings, when the basic issues of sampling error and data dimensionality are not clearly borne in mind.
References
Astúa, D. (2009). Evolution of scapula size and shape in Didelphid Marsupials (Didelphimorphia: Didelphidae). Evolution, 63(9), 2438–2456. https://doi.org/10.1111/j.1558-5646.2009.00720.x.
Baab, K. L. (2016). The role of neurocranial shape in defining the boundaries of an expanded Homo erectus hypodigm. Journal of Human Evolution, 92, 1–21. https://doi.org/10.1016/j.jhevol.2015.11.004.
Benazzi, S., Douka, K., Fornai, C., Bauer, C. C., Kullmer, O., Svoboda, J., et al. (2011). Early dispersal of modern humans in Europe and implications for Neanderthal behaviour. Nature, 479(7374), 525–528. https://doi.org/10.1038/nature10617.
Blackith, R. E., & Reyment, R. A. (1971). Multivariate morphometrics. New York: Academic Press.
Bookstein, F. L. (1991). Morphometric tools for landmark data: Geometry and Biology. New York: Cambridge Univ. Press.
Bookstein, F. L. (1997). Landmark methods for forms without landmarks: Morphometrics of group differences in outline shape. Medical Image Analysis, 1, 225–243.
Bookstein, F. L. (2017). A newly noticed formula enforces fundamental limits on geometric morphometric analyses. Evolutionary Biology, 44(4), 522–541. https://doi.org/10.1007/s11692-017-9424-9.
Bookstein, F. L. (2018). A course in morphometrics for biologists. New York: Cambridge Univ. Press.
Bookstein, F. L. (2019). Pathologies of between-groups principal components analysis in geometric morphometrics. Evolutionary Biology. https://doi.org/10.1101/627448.
Bookstein, F., Schäfer, K., Prossinger, H., Seidler, H., Fieder, M., Stringer, C., et al. (1999). Comparing frontal cranial profiles in archaic and modern Homo by morphometric analysis. The Anatomical Record, 257(6), 217–224. https://doi.org/10.1002/(SICI)1097-0185(19991215)257:6%3c217:AID-AR7%3e3.0.CO;2-W.
Boulesteix, A.-L. (2005). A note on between-group PCA. International Journal of Pure and Applied Mathematics, 19, 359–366.
Cardini, A. (2003). The geometry of the marmot (Rodentia: Sciuridae) mandible: Phylogeny and patterns of morphological evolution. Systematic Biology, 52(2), 186–205. https://doi.org/10.1080/10635150390192807.
Cardini, A. (2018). Integration and modularity in Procrustes shape data: Is there a risk of spurious results? Evolutionary Biology. https://doi.org/10.1007/s11692-018-9463-x.
Cardini, A., & Elton, S. (2007). Sample size and sampling error in geometric morphometric studies of size and shape. Zoomorphology, 126(2), 121–134. https://doi.org/10.1007/s00435-007-0036-2.
Cardini, A., & Elton, S. (2008). Does the skull carry a phylogenetic signal? Evolution and modularity in the guenons. Biological Journal of the Linnean Society, 93(4), 813–834. https://doi.org/10.1111/j.1095-8312.2008.01011.x.
Cardini, A., & Elton, S. (2017). Is there a “Wainer’s rule”? Testing which sex varies most as an example analysis using GueSDat, the free Guenon Skull Database. Hystrix, the Italian Journal of Mammalogy, 28(2), 147–156. https://doi.org/10.4404/hystrix-28.2-12139.
Cardini, A., Jansson, A., & Elton, S. (2007). A geometric morphometric approach to the study of ecogeographical and clinal variation in vervet monkeys. Journal of Biogeography, 34(10), 1663–1678. https://doi.org/10.1111/j.1365-2699.2007.01731.x.
Cardini, A., & O’Higgins, P. (2004). Patterns of morphological evolution in Marmota (Rodentia, Sciuridae): Geometric morphometrics of the cranium in the context of marmot phylogeny, ecology and conservation. Biological Journal of the Linnean Society, 82(3), 385–407. https://doi.org/10.1111/j.1095-8312.2004.00367.x.
Cardini, A., Seetah, K., & Barker, G. (2015). How many specimens do I need? Sampling error in geometric morphometrics: testing the sensitivity of means and variances in simple randomized selection experiments. Zoomorphology, 134(2), 149–163. https://doi.org/10.1007/s00435-015-0253-z.
Chemisquy, M. A., Prevosti, F. J., Martin, G., & Flores, D. A. (2015). Evolution of molar shape in didelphid marsupials (Marsupialia: Didelphidae): Analysis of the influence of ecological factors and phylogenetic legacy. Zoological Journal of the Linnean Society, 173(1), 217–235. https://doi.org/10.1111/zoj.12205.
Chiozzi, G., Bardelli, G., Ricci, M., De Marchi, G., & Cardini, A. (2014). Just another island dwarf? Phenotypic distinctiveness in the poorly known Soemmerring’s Gazelle, Nanger soemmerringii (Cetartiodactyla: Bovidae), of Dahlak Kebir Island. Biological Journal of the Linnean Society, 111(3), 603–620. https://doi.org/10.1111/bij.12239.
Cooke, S. B., & Terhune, C. E. (2015). Form, function, and geometric morphometrics. The Anatomical Record, 298(1), 5–28. https://doi.org/10.1002/ar.23065.
Corti, M., Aguilera, M., & Capanna, E. (2001). Size and shape changes in the skull accompanying speciation of South American spiny rats (Rodentia: Proechimys spp.). Journal of Zoology, 253(4), 537–547. https://doi.org/10.1017/s0952836901000498.
Cucchi, T., Hulme-Beaman, A., Yuan, J., & Dobney, K. (2011). Early Neolithic pig domestication at Jiahu, Henan Province, China: Clues from molar shape analyses using geometric morphometric approaches. Journal of Archaeological Science, 38(1), 11–22. https://doi.org/10.1016/j.jas.2010.07.024.
Culhane, A. C., Perrière, G., Considine, E. C., Cotter, T. G., & Higgins, D. G. (2002). Between-group analysis of microarray data. Bioinformatics, 18(12), 1600–1608. https://doi.org/10.1093/bioinformatics/18.12.1600.
Dapporto, L., Petrocelli, I., & Turillazzi, S. (2011). Incipient morphological castes in Polistes gallicus (Vespidae, Hymenoptera). Zoomorphology, 130(3), 197–201. https://doi.org/10.1007/s00435-011-0130-3.
Domjanic, J., Seidler, H., & Mitteroecker, P. (2015). A combined morphometric analysis of foot form and its association with sex, stature, and body mass. American Journal of Physical Anthropology, 157(4), 582–591. https://doi.org/10.1002/ajpa.22752.
Ferretti, A., Cardini, A., Crampton, J. S., Serpagli, E., Sheets, H. D., & Štorch, P. (2013). Rings without a lord? Enigmatic fossils from the lower Palaeozoic of Bohemia and the Carnic Alps. Lethaia, 46(2), 211–222. https://doi.org/10.1111/let.12004.
Fisher, R. A. (1936). The use of multiple measurements in taxonomic problems. Annals of eugenics, 7(2), 179–188. https://doi.org/10.1111/j.1469-1809.1936.tb02137.x.
Franchini, P., Fruciano, C., Spreitzer, M. L., Jones, J. C., Elmer, K. R., Henning, F., et al. (2014). Genomic architecture of ecologically divergent body shape in a pair of sympatric crater lake cichlid fishes. Molecular Ecology, 23(7), 1828–1845. https://doi.org/10.1111/mec.12590.
Franklin, D., Cardini, A., Flavel, A., & Kuliukas, A. (2013). Estimation of sex from cranial measurements in a Western Australian population. Forensic Science International, 229(1), 158.e151–158.e158. https://doi.org/10.1016/j.forsciint.2013.03.005.
Fruciano, C., Celik, M. A., Butler, K., Dooley, T., Weisbecker, V., & Phillips, M. J. (2017). Sharing is caring? Measurement error and the issues arising from combining 3D morphometric datasets. Ecology and Evolution, 7(17), 7034–7046. https://doi.org/10.1002/ece3.3256.
Fruciano, C., Franchini, P., Raffini, F., Fan, S., & Meyer, A. (2016). Are sympatrically speciating Midas cichlid fish special? Patterns of morphological and genetic variation in the closely related species Archocentrus centrarchus. Ecology and Evolution, 6(12), 4102–4114. https://doi.org/10.1002/ece3.2184.
Fruciano, C., Tigano, C., & Ferrito, V. (2011). Geographical and morphological variation within and between colour phases in Coris julis (L. 1758), a protogynous marine fish. Biological Journal of the Linnean Society, 104(1), 148–162. https://doi.org/10.1111/j.1095-8312.2011.01700.x.
Galimberti, F., Sanvito, S., Vinesi, M. C., & Cardini, A. (2019). Nose-metrics of wild southern elephant seal (Mirounga leonina) males using photogrammetry and geometric morphometry. Journal of Zoological Systematics & Evolutionary Research. https://doi.org/10.1111/jzs.12276.
Gómez-Robles, A., Olejniczak, A. J., Martinón-Torres, M., Prado-Simón, L., & Castro, J. M. B. (2011). Evolutionary novelties and losses in geometric morphometrics: A practical approach through hominin molar morphology. Evolution, 65(6), 1772–1790. https://doi.org/10.1111/j.1558-5646.2011.01244.x.
Gonzalez, P. N., Kristensen, E., Morck, D. W., Boyd, S., & Hallgrímsson, B. (2013). Effects of growth hormone on the ontogenetic allometry of craniofacial bones. Evolution & Development, 15(2), 133–145. https://doi.org/10.1111/ede.12025.
Green, D. J., Sugiura, Y., Seitelman, B. C., & Gunz, P. (2015). Reconciling the convergence of supraspinous fossa shape among hominoids in light of locomotor differences. American Journal of Physical Anthropology, 156(4), 498–510. https://doi.org/10.1002/ajpa.22695.
Gunz, P., & Mitteroecker, P. (2013). Semilandmarks: A method for quantifying curves and surfaces. Hystrix, the Italian journal of mammalogy, 24(1), 103–109. https://doi.org/10.4404/hystrix-24.1-6292.
Gunz, P., Ramsier, M., Kuhrig, M., Hublin, J.-J., & Spoor, F. (2012). The mammalian bony labyrinth reconsidered, introducing a comprehensive geometric morphometric approach. Journal of Anatomy, 220(6), 529–543. https://doi.org/10.1111/j.1469-7580.2012.01493.x.
Hair, J. F., Black, W. C., Babin, B. J., & Anderson, R. E. (2009). Multivariate data analysis (7th ed.). Upper Saddle River: Pearson Prentice Hall.
Hublin, J.-J., Ben-Ncer, A., Bailey, S. E., Freidline, S. E., Neubauer, S., Skinner, M. M., et al. (2017). New fossils from Jebel Irhoud, Morocco and the pan-African origin of Homo sapiens. Nature, 546(7657), 289–292. https://doi.org/10.1038/nature22336.
Ivanović, A., Sotiropoulos, K., Džukić, G., & Kalezić, M. L. (2009). Skull size and shape variation versus molecular phylogeny: A case study of alpine newts (Mesotriton alpestris, Salamandridae) from the Balkan Peninsula. Zoomorphology, 128(2), 157–167. https://doi.org/10.1007/s00435-009-0085-9.
Izenman, A. J. (2008). Modern statistical techniques: regression, classification, and manifold learning. New York: Springer.
Klenovšek, T., & Jojić, V. (2016). Modularity and cranial integration across ontogenetic stages in Martino’s vole, Dinaromys bogdanovi. Contributions to Zoology, 85(3), 275–289. https://doi.org/10.1163/18759866-08503002.
Knigge, R. P., Tocheri, M. W., Orr, C. M., & McNulty, K. P. (2015). Three-dimensional geometric morphometric analysis of talar morphology in extant gorilla taxa from highland and lowland habitats. The Anatomical Record, 298(1), 277–290. https://doi.org/10.1002/ar.23069.
Kovarovic, K., Aiello, L. C., Cardini, A., & Lockwood, C. A. (2011). Discriminant function analyses in archaeology: Are classification rates too good to be true? Journal of Archaeological Science, 38(11), 3006–3018.
Kubiak, B. B., Gutiérrez, E. E., Galiano, D., Maestri, R., & Freitas, T. R. O. (2017). Can niche modeling and geometric morphometrics document competitive exclusion in a pair of subterranean rodents (Genus Ctenomys) with tiny parapatric distributions? Scientific Reports, 7(1), 1–13. https://doi.org/10.1038/s41598-017-16243-2.
Mahalanobis, P. C. (1936). On the generalized distance in statistics. Proceedings National Institute of Science, India, 2(1), 49–55.
Mitteroecker, P., & Bookstein, F. (2011). Linear discrimination, ordination, and the visualization of selection gradients in modern morphometrics. Evolutionary Biology, 38(1), 100–114. https://doi.org/10.1007/s11692-011-9109-8.
Mitteroecker, P., Gunz, P., & Bookstein, F. L. (2005). Heterochrony and geometric morphometrics: A comparison of cranial growth in Pan paniscus versus Pan troglodytes. Evolution & Development, 7(3), 244–258. https://doi.org/10.1111/j.1525-142X.2005.05027.x.
Neubauer, S., Gunz, P., Leakey, L., Leakey, M., Hublin, J.-J., & Spoor, F. (2018). Reconstruction, endocranial form and taxonomic affinity of the early Homo calvaria KNM-ER 42700. Journal of Human Evolution, 121, 25–39. https://doi.org/10.1016/j.jhevol.2018.04.005.
Oxnard, C., & Higgins, P. (2011). Biology clearly needs morphometrics. Does morphometrics need biology? Biological Theory, 4(1), 84–97. https://doi.org/10.1162/biot.2009.4.1.84.
Pallares, L. F., Turner, L. M., & Tautz, D. (2016). Craniofacial shape transition across the house mouse hybrid zone: Implications for the genetic architecture and evolution of between-species differences. Development Genes and Evolution, 226(3), 173–186. https://doi.org/10.1007/s00427-016-0550-7.
Ritzman, T. B., Terhune, C. E., Gunz, P., & Robinson, C. A. (2016). Mandibular ramus shape of Australopithecus sediba suggests a single variable species. Journal of Human Evolution, 100, 54–64. https://doi.org/10.1016/j.jhevol.2016.09.002.
Rohlf, F. J. (2015). The tps series of software. Hystrix, the Italian Journal of Mammalogy, 26, 1-4. https://doi.org/10.4404/hystrix-26.1-11264.
Rohlf, F. J., & Slice, D. (1990). Extensions of the Procrustes method for the optimal superimposition of landmarks. Systematic Zoology, 39(1), 40–59. https://doi.org/10.2307/2992207.
Sanfilippo, P. G., Cardini, A., Sigal, I. A., Ruddle, J. B., Chua, B. E., Hewitt, A. W., et al. (2010). A geometric morphometric assessment of the optic cup in glaucoma. Experimental Eye Research, 91(3), 405–414. https://doi.org/10.1016/j.exer.2010.06.014.
Sansalone, G., Colangelo, P., Kotsakis, T., Loy, A., Castiglia, R., Bannikova, A. A., et al. (2018). Influence of evolutionary allometry on rates of morphological evolution and disparity in strictly subterranean Moles (Talpinae, Talpidae, Lipotyphla, Mammalia). Journal of Mammalian Evolution, 25(1), 1–14. https://doi.org/10.1007/s10914-016-9370-9.
Schlager, S. (2017). Morpho and Rvcg—Shape analysis in R. In G. Zheng, S. Li, & G. Szekely (Eds.), Statistical shape and deformation analysis (pp. 217–256). New York: Academic Press.
Schlager, S., & Rüdell, A. (2015). Analysis of the human osseous nasal shape—Population differences and sexual dimorphism. American Journal of Physical Anthropology, 157(4), 571–581. https://doi.org/10.1002/ajpa.22749.
Seetah, T. K., Cardini, A., & Miracle, P. T. (2012). Can morphospace shed light on cave bear spatial-temporal variation? Population dynamics of Ursus spelaeus from Romualdova pećina and Vindija, (Croatia). Journal of Archaeological Science, 39(2), 500–510. https://doi.org/10.1016/j.jas.2011.10.005.
Serb, J. M., Sherratt, E., Alejandrino, A., & Adams, D. C. (2017). Phylogenetic convergence and multiple shell shape optima for gliding scallops (Bivalvia: Pectinidae). Journal of Evolutionary Biology, 30(9), 1736–1747. https://doi.org/10.1111/jeb.13137.
Siberchicot, A., Julien-Laferrière, A., Dufour, A.-B., Thioulouse, J., & Dray, S. (2017). adegraphics: An s4 lattice-based package for the representation of multivariate data. The R Journal, 9(2), 198–212. https://doi.org/10.32614/RJ-2017-042.
Skinner, M. M., Gunz, P., Wood, B. A., & Hublin, J. J. (2009). How many landmarks? Assessing the classification accuracy of Pan lower molars using a geometric morphometric analysis of the occlusal basin as seen at the enamel-dentine junction. In T. Koppe (Ed.), Comparative dental morphology. Basel: Karger Publishers. https://doi.org/10.1159/000242385.
Slice, D. E. (2005). Modern morphometrics. In D. E. Slice (Ed.), Modern morphometrics in physical anthropology (pp. 1–45). Boston, MA: Springer.
Souto-Lima, R. B., & Millien, V. (2014). The influence of environmental factors on the morphology of red-backed voles Myodes gapperi (Rodentia, Arvicolinae) in Québec and western Labrador. Biological Journal of the Linnean Society, 112(1), 204–218. https://doi.org/10.1111/bij.12263.
Torres-Tamayo, N., García-Martínez, D., Zlolniski, S. L., Torres-Sánchez, I., García-Río, F., & Bastir, M. (2018). 3D analysis of sexual dimorphism in size, shape and breathing kinematics of human lungs. Journal of Anatomy, 232(2), 227–237. https://doi.org/10.1111/joa.12743.
Watanabe, A. (2018). How many landmarks are enough to characterize shape and size variation? PLoS ONE, 13(6), e0198341. https://doi.org/10.1371/journal.pone.0198341.
Yendle, P. W., & MacFie, H. J. (1989). Discriminant principal components analysis. Journal of Chemometrics, 3(4), 589–600. https://doi.org/10.1002/cem.1180030407.
Acknowledgements
We are very grateful to Jessica Grisenti, who carefully collected the marmot data for her undergraduate thesis and gave AC permission to use them. The authors appreciate the most helpful comments of Julien Claude who reviewed this paper.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflicts of interest.
Additional information
The paper is dedicated to the memories of Nicola Saino (1961–2019) and Dennis Slice (1958–2019). Nicola was one of the greatest Italian ethologists, Professor of Animal Behaviour at the University of Milan, and extraordinary ornithologist: AC will always remember with fondness the day Nicola introduced him, and other biology students, to the wonders of birdwatching; he will also never forget his brilliant example as a teacher and researcher; and he will greatly miss the passionate fights, with him, over methods. Dennis Slice Professor in the Dept. of Scientific Computing, The Florida State University was an Evolutionary biologist and Ecologist who made major contributions to morphometrics through his scientific and software contributions, by maintaining and moderating the MORPHMET discussion group and by being a tireless supporter and educator of students and colleagues. For his contributions, he was awarded the Rohlf Medal for Excellence in Morphometric Methods and Applications in 2017. Not only was he a tireless advocate of his field but he was a wonderful colleague, always available and always thoughtful. We will miss him greatly as both a scientist and a colleague.
Rights and permissions
About this article
Cite this article
Cardini, A., O’Higgins, P. & Rohlf, F.J. Seeing Distinct Groups Where There are None: Spurious Patterns from Between-Group PCA. Evol Biol 46, 303–316 (2019). https://doi.org/10.1007/s11692-019-09487-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11692-019-09487-5