1 Introduction

Fractional factorial designs are widely used in many scientific investigations because they provide a systematic and statistically valid strategy for studying how multiple factors impact a response variable through main effects and interactions. When several factors are to be tested, often the experimenter does not know which factors have important interactions. Instead, the experimenter will need to perform model selection after conducting the experiment to identify important interactions. Generally this process will involve fitting different models under consideration and examining statistical significance of the interaction terms. Some techniques have been developed concerning finding efficient fractional factorial plans for this purpose. There is a rich literature on identification and discrimination to find the model best describing the data [18,19,20].

Consider a design with m factors of interest. Following Ghosh and Chowdhury [7], consider the following class of s candidate models for describing the relationship between p \((\le m)\) of the m factors and the \(n\times 1\) vector of observations \({\varvec{y}}\),

$$\begin{aligned}&E\left( {\varvec{y}}\right) = \beta _0{\varvec{j}}_n + {\varvec{X}}_1 \varvec{\beta }_1 + {\varvec{X}}_{2}^{(i)} \varvec{\beta }_{2i}, \quad i = 1, \ldots , s \nonumber \\&\text{Var}({\varvec{y}}) = \sigma ^2 {\varvec{I}}, \end{aligned}$$
(1)

where n is the number of runs, \(\beta _0\) is the general mean, \({\varvec{j}}_n\) is a vector of ones, \(\varvec{\beta }_1\) is the vector of p main effects that are common in all s models. The other parameters, \(\varvec{\beta }_{2i}\), are specific for the ith model and hence \(\varvec{\beta }_{2i} \ne \varvec{\beta }_{2i^{\prime }}\) for \(i \ne i^{\prime },\) \(i = 1, \dots , s\). We call these parameters “uncommon parameters.” The design matrices \({\varvec{X}}_1\) and \({\varvec{X}}_{2}^{(i)}\) correspond to the main effects and ith set of 2-factor interactions, respectively. Following Ghosh and Flores [8] and Ghosh and Chowdhury [7] we consider the situation of \(p=m\) for generating A-ComVar designs using our proposed approach, described in Sect. 4. However, we consider models with \(p\le m\) cases in our examples comparing the performance of our proposed designs with some popular designs from the literature.

Under the above setup, model selection consists of identifying the correct i from the s candidate models. This process is complicated by the fact that the variance estimates for the uncommon parameters are generally not the same, which can pre-bias the experiment toward identifying certain interactions as significant over others, i.e., making some i more likely to be selected than others regardless of the true underlying model. To address this issue, Ghosh and Flores [8] introduced the notion of common variance designs for a single uncommon parameter. These designs estimate the uncommon parameter in all models with equal variance, which is desirable in the absence of any a priori information about the true model. Ghosh and Chowdhury [7] generalized this concept of common variance to k \((k\ge 1)\) uncommon parameters in each model in the class. Under the situation of \(k > 1\), Ghosh and Chowdhury [7] defined a common variance design to be the one satisfying \(|{\varvec{X}}^{(i) \prime } {\varvec{X}}^{(i)}|\) to be a constant, for all i, \({\varvec{X}}^{(i)} = \left( {\varvec{j}}_n, {\varvec{X}}_1, {\varvec{X}}_2^{(i)} \right) \).

The concept of variance balancedness is not totally new. Different types of “variance-balanced designs” estimating all or some of the treatment contrasts with identical variance were developed by Calvin [3], Cheng [5], Gupta and Jones [10], Hedayat and Stufken [11], Khatri [13], Mukerjee and Kageyama [17], among others.

While common variance designs have been identified for two- and three-level factorial experiments with a single 2-factor interaction [7, 8], it remains to develop a method which can find them for general number of factors and interactions. To date, these designs have been found using exhaustive searches (Table 1), which becomes prohibitively expensive as the number of factors and runs increases. This leads us to introduce approximate common variance (A-ComVar) designs, which relax the requirement that the variance of the uncommon parameters is exactly equal. We introduce an objective function that allows us to rank designs under consideration, and we develop a genetic algorithm for searching for these designs. Moreover, we investigate the performance of both common variance and A-ComVar designs for model selection using the adaptive lasso regression technique in simulation [12]. We find comparable performance of common variance and A-ComVar designs to several other designs for model discrimination (Table 2), which further demonstrates the usefulness of designs that prioritize having a similar variance for the uncommon parameters in the model.

Table 1 Complete search results for finding common variance designs for \(3^3\) factorial experiments
Table 2 Description of the designs used for Example 3

The rest of the article is organized as follows. In Sect. 2 we present the current state of knowledge for both two-level and three-level common variance designs. For three-level designs we also present the exhaustive search result for \(m=3\). In Sect. 3 we introduce our numerical approach for finding A-ComVar designs. In Sect. 4 we conduct extensive studies to both (1) examine our numerical approach’s ability to find A-ComVar designs as we increase the number of factors and number of interactions in the model and (2) compare these A-ComVar designs to potential competitor designs from the literature. Finally, Sect. 5 contains some discussion of the results and some future directions for our work. “Appendix 1” contains an illustration of our genetic algorithm, and “Appendix 2” contains Tables corresponding to all results.

2 Common Variance Designs

2.1 Two-Level Designs

The term “common variance” for the class of variance-balanced designs was first introduced in Ghosh and Flores [8]. As more stringent criteria, the authors also introduced the concept of optimum common variance (OPTCV), which is satisfied by designs having the smallest value of common variance in a class of common variance designs with \(p \le m\) factors and n runs. Several characterizations of common variance and optimal common variance designs were presented that provide efficient ways for checking the common variance or OPTCV property of a given design. These characterizations were obtained in terms of the projection matrix, eigenvalues of the model matrix, balancedness, and orthogonal properties of the designs. In Corollary 1 of Ghosh and Flores [8], they stated one sufficient condition of common variance designs in terms of equality of the vectors of eigenvalues of \({\varvec{X}}^{(i)\prime } {\varvec{X}}^{(i)}\), \({\varvec{X}}^{(i)} = \left( {\varvec{j}}_ n, {\varvec{X}}_1, {\varvec{X}}_2^{(i)} \right) \), for all i. We present one design in Table 3 from Ghosh and Flores [8] for \(m=5\) and \(n=12\) that has identical vectors of eigenvalues for all i. In Sect. 4.3 we compare the performance of this particular design with that of a Plackett–Burman design for model selection to demonstrate further usefulness of such designs.

In their work, Ghosh and Flores [8] presented several general series of designs with the common variance property. For example, they identified two fold-over designs with the common variance property with all m factors of the design and \(n=2m\) and \(n=2m+2\) runs, respectively:

$$\begin{aligned}&d_m^{(2m)}= \left[ {\begin{array}{c} 2{\varvec{I}}_m - {\varvec{J}}_m\\ -2{\varvec{I}}_m + {\varvec{J}}_m \\ \end{array} } \right] .\\&d_m^{(2m+2)}= \left[ {\begin{array}{c} {\varvec{j}}_m^{\prime } \\ -{\varvec{j}}_m^{\prime } \\ 2{\varvec{I}}_m - {\varvec{J}}_m\\ -2{\varvec{I}}_m + {\varvec{J}}_m \\ \end{array} } \right] . \end{aligned}$$

As reported in Ghosh and Flores [8], both of these designs are balanced arrays of full strength and orthogonal arrays of strength 1, for all m. Moreover, the design \({\varvec{d}}^{(2m)}_m\) is OPTCV for \(m=4\) and \({\varvec{d}}^{(2m+2)}_m\) is OPTCV for \(m=3\).

2.2 Three-Level Designs

Ghosh and Chowdhury [7] presented common variance designs for \(3^m\) fractional factorial experiments. Consider the following model for a \(3^m\) factorial experiment, with one 2-factor interaction effect in the model, i.e., \(k=1\):

$$\begin{aligned} E\left( {\varvec{y}}\right) = \beta _0{\varvec{j}}_n + {\varvec{X}}_1 \varvec{\beta }_1 + {\varvec{X}}_{2}^{(i)} \beta _{2i}, \quad \text{Var}({\varvec{y}}) = \sigma ^2 {\mathbf{I}}. \end{aligned}$$

A design for such an experiment would have the common variance property iff \(\frac{\mathrm{Var}({\hat{\beta }}_{2i})}{\sigma ^2}\) is constant for all \(i=1, \dots , 4\left( {\begin{array}{c}m\\ 2\end{array}}\right) \), for the situation \(p=m\).

Ghosh and Chowdhury [7] presented two general series of \(3^m\) fractional factorial common variance designs \(d_1\) and \(d_2\) with n runs. The design \(d_1\) has a common variance value given by \(\frac{\mathrm{Var}\left( {\hat{\beta }}_2^{(i)}\right) }{\sigma ^2} = \frac{2-m+m^2}{9}\), for \(m \ge 2\) and \(n=2m+2\) runs, while design \(d_2\) has a common variance value given by \(\frac{\mathrm{Var}\left( {\hat{\beta }}_2^{(i)}\right) }{\sigma ^2} = \frac{m}{9(m-2)}\), for \(m \ge 3\) and \(n=3m\). Also, the design \(d_1\) is efficient common variance (ECV, as termed in Ghosh and Chowdhury [7]) design for \(m=2\), and design \(d_2\) is ECV for \(m=3\).

Ghosh and Chowdhury [7] also presented several sufficient conditions for general fractional factorial designs to have the common variance property, including the special case for \(3^m\) designs in terms of the projection matrix of the design and the columns of 2-factor interaction. For example, a design is common variance if (1) \({{\varvec{P}}}{{\varvec{X}}}_2^{(i_1)} = {{\varvec{P}}}{{\varvec{X}}}_2^{(i_2)}\), for \(i_1, i_2 \in \{1, \dots , s\}\), where \({\varvec{P}}\) is the projection matrix defined as \({\varvec{I}}_n - {\varvec{X}}_1 \left( {\varvec{X}}_1^{\prime } {\varvec{X}}_1\right) ^{-1} {\varvec{X}}_1^{\prime }\), and \({\varvec{X}}_1\) contains the columns corresponding to the general mean and main effects from the model matrix \({\varvec{X}}^{(i)} = \left( {\varvec{j}}_n, {\varvec{X}}_1, {\varvec{X}}_2^{(i)}\right) \), and \({\varvec{X}}_2^{(i)}\) corresponds to the \(i^{th}\) 2-factor interaction. Another set of sufficient conditions for having common variance is, for, \(i_1, i_2 \in \{1, \dots , s\}\), (1) \(\left( {\varvec{X}}_2^{(i_1)} \pm {\varvec{X}}_2^{(i_2)}\right) \) belongs to the column space of \({\varvec{X}}_1\) and (2) \({\varvec{X}}_2^{(i_2)} = {\varvec{F}} {\varvec{X}}_2^{(i_1)}\) holds, where the permutation matrix \({\varvec{F}}\) obtained from the identity matrix satisfies \({\varvec{F}}^{\prime } {\varvec{P}} {\varvec{F}} = {\varvec{P}}\).

For \(3^3\) fractional factorial experiment, Chowdhury [6] conducted a complete search of common variance designs for \(n=8\) to \(n=27\), since \(n=8\) is the minimum number of runs needed to estimate all the parameters considering all 3 factors are present in the model (one general mean, 6 main effects, one 2-factor interaction effect). The results of this search are presented in Table 1. The complete search revealed that common variance designs only exist for \(n=8,9,10,11\) for \(3^3\) factorial experiments. For each of the runs multiple groups of common variance designs were obtained, having different common variance values, among which 32 designs for \(n=11\); 48 designs for \(n=10\); 8256 designs for \(n=9\); and 9600 designs for \(n=8\), are the efficient common variance designs giving the minimum value of common variance in the respective classes.

3 Identifying Common Variance Designs

3.1 Challenges in Numerically Identifying Common Variance Designs

Ghosh and Flores [8] and Ghosh and Chowdhury [7] presented some general series of designs satisfying the common variance property for two- and three-level factorial experiments obtained via exhaustive searches of the design space. Such searches become extremely computationally challenging as the number of factors increases. For example, for a \(3^3\) factorial experiment with one 2-factor interaction (\(k=1\)) the possible set of candidate designs with 8 runs is \(\left( {\begin{array}{c}27\\ 8\end{array}}\right) =2{,}220{,}075\), with 9 runs is \(\left( {\begin{array}{c}27\\ 9\end{array}}\right) =4{,}686{,}825\), with 10 runs is \(\left( {\begin{array}{c}27\\ 10\end{array}}\right) =8{,}436{,}285\), and so on. For a \(3^4\) factorial experiment, the cardinality of this set increases to \(\left( {\begin{array}{c}81\\ 10\end{array}}\right) =1.878392\times 10^{12}\), even for the designs with the smallest possible number of runs. This rapid growth in the size of the search space makes exhaustive searches for common variance designs impossible for anything but small design problems.

In light of the difficulty in finding common variance designs, we introduce a class of approximate common variance (A-ComVar) designs. Instead of having exactly equal variance for the uncommon parameters for the s models under consideration, A-ComVar designs try to maximize the ratio of the minimum variance to the maximum variance. In doing so, they contain common variance designs as a sub-case where the minimum variance is exactly equal to the maximum variance. In relaxing the requirement that the variances are exactly equal, we are able to develop an objective function and algorithm for identifying these A-ComVar designs without performing an exhaustive search.

3.2 Proposed Algorithm: Genetic Algorithm for Finding A-ComVar Designs

In this section we propose to use a genetic algorithm to identify A-ComVar designs. We start by defining an objective function that seeks to quantify our goal. Denote the variance of the interaction effect for the ith model as \(\sigma _{\beta _{2i}}^2\) and let \(\overline{\sigma _{\beta _{2}}^2} = \frac{1}{s} \sum _{i} \sigma _{\beta _{2i}}^2\). The objective function for designs that discriminate between models with a single interaction term (\(k=1\)) is:

$$\begin{aligned} f(d; \phi ) = \frac{1 / \overline{\sigma _{\beta _{2}}^2} }{1 + \phi \times \sum _{i=1}^{s} (\sigma _{\beta _{2i}}^2 - \overline{\sigma _{\beta _{2}}^2} )^2 }, \end{aligned}$$
(2)

where \(\sigma _{\beta _{2i}}^2\) is replaced by the determinant of the lower-right \(k \times k\) sub-matrix of the inverse of the Fisher information matrix for \(k > 1\), which bears some similarity to the idea behind D-optimal design of experiments. The value of the objective function increases as the variance of the estimates decreases through the numerator, encouraging designs with small variances for the interaction terms. However, this value is also strongly penalized toward zero as the individual model variances move away from the average model variance. The strength of this penalty is controlled by the tuning parameter \(\phi \), which we recommend setting to a very large value. In our experiments we found \(\phi = 10 \times 10^{13}\) to be adequate. The \(\phi \) parameter is just to force differences in variance across models under consideration to “cost” more than the potential variance improvement from a design under some subset of those models, and thus setting it to any suitable large value should suffice. Taken together the numerator allows us to differentiate between designs with common variance to select the better one, and the denominator encourages common variance designs by penalizing differing variance under alternative models under consideration.

This maximization approach will prefer A-ComVar designs with exactly common variance. Of course, in many experimental situations a common variance design may not exist. For example in the exhaustive search, Chowdhury [6] found that common variance designs did not exist for \(3^3\) experiments for 13 runs. This leads us to the principal advantage of our approach: when a common variance design does not exist we can still find designs with variance that is as close as possible to being equal. To assess the quality of an A-ComVar design, we define the A-ComVar ratio

$$\begin{aligned} r_{\mathrm{ACV}} = \frac{\min \nolimits _{i}\left\{ \mathrm{var}\left( {\hat{\beta }}_{2i}\right) \right\} }{\max \nolimits _{i}\left\{ \mathrm{var}\left( {\hat{\beta }}_{2i}\right) \right\} }. \end{aligned}$$
(3)

Clearly when a design has common variance, \(r_{\mathrm{ACV}} = 1\). When a design does not have common variance, \(r_{\mathrm{ACV}}\) gives us an idea of how far we are from common variance. For example, if \(r_{\mathrm{ACV}} = 0.5\) then we know that among the models under consideration, the largest variance of interaction terms is twice that of the smallest. This knowledge can hopefully help inform model selection.

Any off-the-shelf optimization algorithm could be used to try to maximize this objective function. We have chosen to use a genetic algorithm, as is common in the design literature [15, 16]. Genetic algorithms are optimization techniques mimicking Darwin’s idea of natural selection and survival of the fittest. This search expects that a good candidate solution will provide good offspring and imitates the way that chromosomes crossover and mutate when reproducing. Here, each chromosome is a design, and the fitness of a chromosome is determined by the corresponding objective function value. At each iteration the worst chromosomes are replaced with offspring generated by combining the settings from two better chromosomes, along with some small probability of a mutation. In the context of our problem, a mutation corresponds to randomly changing the settings for one of the factors in one of the runs. The algorithm terminates when either the maximum number of iterations has been reached, or a design with common variance has been found. The steps in our genetic algorithm are outlined in Algorithm 1, and “Appendix 1” provides a detailed example of how our algorithm is used.

The genetic algorithm requires the user to specify the mutation probability, the number of chromosomes to replace at each iteration, and the maximum number of iterations. Our experience with the algorithm suggests using a small mutation probability to encourage only one or two mutations each time a new chromosome is created. Similarly, we have found replacing two chromosomes at each iteration to work for our purposes, and so throughout the remainder of this paper we fix this tuning parameter at two. Finally, we generally use a maximum of 10,000 iterations, although the algorithm is quite fast and this number can easily be increased if needed. Our algorithm is implemented in Julia version 1.0.2 and is available for download from the author’s website.

figure a

4 Numerical Examples

4.1 Example 1: Designs with One 2-Factor Interaction

We conducted a series of experiments to investigate the ability of our approach to find A-ComVar designs and to gain a better understanding of when common variance designs can be found. We started by examining designs with a single 2-factor interaction. We consider \(2^{m_1}\) and \(3^{m_2}\) experiments, with \(m_1 = 4, \ldots , 9\) and \(m_2 = 3, \ldots , 6\) and consider the situation where all factors are present in the model as main effects. For the \(2^{m_1}\) experiments, we considered run sizes of \(n_{m_1} = m_1 + 2, \ldots , m_1 + 11\), and for the \(3^{m_2}\) experiments, we considered run sizes of \(n_{m_2} = 2m_1 + 2, \ldots , 2m_1 + 11\). For each combination of settings, we ran our genetic algorithm 100 times and stored the \(r_{\mathrm{ACV}}\) results. The tuning parameters used were a mutation probability of 0.05 and a maximum of 10,000 iterations.

Figure 1 displays the results for the \(2^{m_1}\) cases, and Fig. 2 displays the results for the \(3^{m_2}\) cases. We first note that our results are consistent with the findings of Ghosh and Chowdhury [7], who used exhaustive searches to identify common variance designs. For example, Ghosh and Chowdhury [7] found that common variance designs exist for \(3^3\) designs with 8 runs, which agrees with the boxplots in the first panel of Fig. 2. This supports our use of the genetic algorithm approach with the objective function described above. Furthermore, in cases where the common variance designs either do not exist or could not be found, our approach was able to find designs that attempt to get as close as possible to common variance. For example, it is known from exhaustive searches that no common variance design exists for a \(3^3\) experiment with 12 runs. However, the proposed approach was able to find designs where the smallest variance was greater than 0.8 times the largest variance, indicating that the design is quite close to having the common variance property.

Fig. 1
figure 1

Ratios \(r_{\mathrm{ACV}}\) for the \(2^{m_1}\) case across 100 replicates for each experimental setting

Fig. 2
figure 2

Ratios \(r_{\mathrm{ACV}}\) for the \(3^{m_2}\) case across 100 replicates for each experimental setting

Fig. 3
figure 3

Ratios \(r_{\mathrm{ACV}}\) for the \(2^{m_1}\) case across 100 replicates for each experimental setting with \(k=2\)

4.2 Example 2: Designs with Two 2-Factor Interactions

For designs with multiple 2-factor interactions (i.e., \(k > 1\)), we generalize the objective function in (2) by replacing \(\text{var}({\hat{\beta }}_{2i})\) with the determinant of the block of the inverse of the Fisher information matrix corresponding to the interaction terms.

To demonstrate the approach, we conducted another experiment with two 2-factor interactions (i.e., \(k=2\)). We consider \(2^{m_3}\) experiments, with \(m_3 = 4, \ldots , 7\), and assume \(p=m_3\). We considered run sizes of \(n_{m_1} = m_3 + 6, \ldots , m_3 + 12\). For each combination of settings, we ran our genetic algorithm 100 times and stored the \(r_{\mathrm{ACV}}\) results. The tuning parameters used were a mutation probability of 0.05 and a maximum of 10,000 iterations.

Figure 3 shows the results. As before, we can see that in many cases the genetic algorithm is able to find common variance designs. In cases where common variance designs cannot be found, the approach is often able to identify a design resulting in relatively close to common variance.

4.3 Example 3: Common Variance Design and A-ComVar Designs for Model Selection

We next perform a series of studies to demonstrate the advantages of pursuing A-ComVar designs. We do this by considering data generated from a variety of true models and testing whether a model selection procedure is able to identify the true model using observations collected using the designs under consideration. We used the adaptive lasso [12] to fit the model. We chose the adaptive lasso method of Kane and Mandal [12] because they showed that this technique is suitable for identifying the correct model for designs with complex aliasing and that it outperforms other popular variable selection methods including the Dantzig selector [4], LARS [21], and the nonnegative Garotte estimator [2, 22].

Our procedure is as follows. For a model with p active main effects, let \(F_1, \ldots , F_p\) denote the active factors, which are selected at random from the set of all factors of the designs at each replication. The corresponding effects, \(\beta _1, \ldots , \beta _p\), as well as any interaction effects, are set to be either “big” or “small,” where “big” effects are drawn from a U(1.5, 2.5) distribution and “small” effects from a U(0.1, 0.3). Finally, the error standard deviation, \(\sigma \), is chosen, completing the specification of the true underlying model. The total number of different models, effect sizes, and error standard deviations considered can be found in any of Tables 8, 9, 10, 11, 12, 13, and 14. The first column in each table corresponds to the true model under consideration, and the second column gives information about the strengths of the active effects (b—“big” and s—“small”). For example, row 25 of Table 10 corresponds to a model with three active main effects (\(F_1\), \(F_2\), and \(F_3\)) as well as one active interaction (\(F_1 F_3\)). Here, the second column tells us that \(F_1\) and \(F_2\) have “big” effects and \(F_3\) and \(F_1 F_3\) have “small” effects.

Next, for each design under consideration, a data set is generated from the true underlying model using the randomly selected factors of the design. A model is fit to this data set using the adaptive lasso, and we measure whether or not the true underlying model was identified. This process is then repeated 100 times for the same set of true active coefficients, and we store the percentage of the times the correct model was identified.

For each model, design, and error standard deviation under consideration, this process of randomly selecting active factors in the model, generating observations from the design, and measuring how often the correct model is identified is repeated 50 times, resulting in 50 replicates per combination of settings. Here each replicate is a measurement of the percentage of times the data obtained using the design was able to correctly identify the true underlying model. Table 2 displays a list of the model comparisons we made. In Tables 8, 9, 10, 11, 12, 13, and 14 we report the average percentage of times (over 50 replications) the correct model was identified by the respective designs (Tables 3, 4, 5, 6).

Table 3 Design \({\varvec{d}}_5^{(12)}\) with common variance for 5 factors and 12 runs (left) and Plackett–Burman design with 11 factors and 12 runs (right)
Table 4 (1) \(D^1\)(left): two-level A-ComVar design for \(k=1\) with \(m=5\) and \(n=12\), (2) \(D^2\) (middle): three-level A-ComVar design for \(k=1\) with \(m=4\) and \(n=20\) and (3) \(D^3\) (right): three-level A-ComVar design for \(k=1\), \(m=7\), \(n=18\), used in Example 3
Table 5 (1) \(D^4\)(left): two-level Bayes optimal design with \(m=5\) and \(n=12\) from Bingham and Chipman [1], (2) \(D^5\) (middle): two-level design with \(m=5\) and \(n=12\) from Li and Nachesheim [14] and (3) \(D^6\) (right): two-level design with \(m=5\), \(n=12\) from Ghosh and Tian [9], used in Example 3
Table 6 (1) \(D^7\)(left): central composite design (CCD) with \(m=3\) and \(n=20\), (2) \(D^8\) (right): three-level orthogonal main effect plan (OME) with \(m=7\) and \(n=18\) used in Example 3

The results with model \(\times \) variance \(\times \) design breakdown can be found in Tables 8, 9, 10, 11, 12, 13, and 14 in “Appendix 2.” Figures 4, 5, and 6 present boxplots of the results for each standard deviation level, stratified by the number of interactions in the model (0, 1, or 2). From the figures we can see that while the common variance design outperforms Plackett–Burman design for all three types of models, a few general patterns are observed for A-ComVar designs. First, when the model contains only main effects and no interactions, the A-ComVar designs generally perform about as well as the competitor designs. This is important, as it suggests that there is not a strong disadvantage to seeking such designs in practice. Next, examining the boxplots for the models with one and two interactions, we can see that the A-ComVar designs generally outperform the other designs, especially for the models with two interactions. The exception to this pattern is the Ghosh and Tian design, which seems to outperform the A-ComVar design in several cases. This is likely because the Ghosh and Tian [9] design is optimal w.r.t all six standard optimality criteria, and thus it is hard to beat its performance. However, designs of this quality cannot always be obtained for arbitrary numbers of factors or runs; thus, one advantage of our numerical approach is that it can be used for cases where such designs cannot be obtained via exhaustive search or by using theoretical results.

Fig. 4
figure 4

Boxplots of the results for each standard deviation level, stratified by the number of interactions in the model (0, 1, or 2 from left to right) for model selection performance of two-level common variance design \(d_5^{(12)}\) design with 5 factors and 12 runs presented in Table 2 with PB design presented in Table 3. Here represents the common variance design \(d_5^{(12)}\) and represents Plackett–Burman design

Fig. 5
figure 5

Boxplots of the results for each standard deviation level, stratified by the number of interactions in the model (0, 1, or 2 from left to right) for model selection performance of two-level A-ComVar design with 5 factors and 12 runs with other designs. Here represents our A-ComVar design with 2 levels, represents Plackett–Burman design, represents a two-level design with 5 factors and 12 runs from Ghosh and Tian [9], represents a two-level Bayes optimal design with 5 factors and 12 runs from Bingham and Chipman [1], represents a two-level design with 5 factors and 12 runs from Li and Nachtsheim [14]

Fig. 6
figure 6

Boxplots of the results for each standard deviation level, stratified by the number of interactions in the model (0, 1, or 2 from left to right) for model selection performance of three-level A-ComVar designs with central composite and orthogonal main effect designs. Here represents our three-level A-ComVar design with 4 factors and 20 runs, represents a central composite design with 3 factors and 20 runs, represents our three-level A-ComVar design with 7 factors and 18 runs, represents a orthogonal main effect plan with 7 factors and 18 runs

5 Discussion

In this work we introduced A-ComVar designs, an extension of common variance designs. Our proposed approach addresses the difficulties associated with finding common variance designs via exhaustive search. Through several examples, we demonstrated that the proposed algorithmic approach allows us to quickly find common variance designs that overlap with those known in the literature. Furthermore, in cases where common variance designs do not exist or cannot be found, our approach allows identification of designs with close to common variance. Comparisons to a Plackett–Burman design and several other standard optimal designs from the literature demonstrated that such designs perform quite well in practice, and that in many cases these A-ComVar designs perform as well as common variance designs.

There are several avenues here for future work. First, we considered only the cases with two-level and three-level factors. Future work could consider finding A-ComVar designs with mixed−level factors. Second, we utilized a genetic algorithm to find these designs. There are numerous other optimization approaches that could be used to maximize the objective function in (2). In some cases, these other approaches may succeed in finding designs with a better ratio of minimum to maximum variance of the uncommon parameters. Third, there is another approach to finding common variance designs through hierarchical designs [6]. These designs are found by identifying a common variance design for a smaller number of runs and then adding runs while trying to preserve the common variance property. It is possible that a similar idea could be developed for A-ComVar designs. Finally, future work could study the types of A-ComVar designs that can be found when the number of interactions in the model increases beyond two.