1 Introduction

Engineering systems design often requires the coordination of numerous computationally-intensive multidisciplinary analyses to account for their interactions and ensure overall system consistency. Martins and Lambe provide an excellent overview of MDO coordination methods in Martins and Lambe (2013). In this paper, we consider the non-hierarchical formulation of Analytical Target Cascading (ATC). ATC is a methodology developed originally for managing component requirements in hierarchically decomposed optimal system design problems (Michelena et al. 1999; Kim 2001). It deals with the consistency constraints that arise from a hierarchical object-based decomposition of a system design problem by coordinating target-response pairs among hierarchy levels. Its efficiency for solving decomposition-based optimal system design problems has been demonstrated in several studies, see, e.g., Kim et al. (2002, 2003), Kokkolaras et al. (2002, 2004) and Allison et al. (2005). The system cascades design targets for coupling and/or shared variables to subsystems, the subsystems to components and so on. Optimization subproblems are formulated for each element in the hierarchy, and solved iteratively according to a coordination strategy until consistency has been achieved. This process has been shown to be convergent under standard convexity and continuity assumptions (Michelena et al. 2003; Tosserams et al. 2006; Kim et al. 2006).

The ATC methodology was extended in Tosserams et al. (2010) by means of a non-hierarchical formulation that enables the coordination of general distributed multidisciplinary design optimization (MDO) problems. This extension allows direct treatment of functional dependencies among all components of a decomposed design problem (see Fig. 1). The formulations presented in Tosserams et al. (2010) also introduce system-wide functions to manage design attributes shared among all elements (e.g., weight), reducing thus the number of target-response pairs that must be coordinated. Finally, non-hierarchical analytical target cascading facilitates concurrent processing of all (as opposed to same-level only) optimization subproblems within an ATC coordination iteration. Non-hierarchical coordination introduces local copies of variables that link and/or are shared by elements. Every optimization subproblem is solved with respect to these local copies of variables. Consistency is ensured using Augmented Lagrangian Coordination (ALC) with quadratic penalty functions (Tosserams et al. 2008). The linear and quadratic weights of the penalty functions are updated using the method of multipliers (formulation details are given in Section 2).

Fig. 1
figure 1

Analysis flow example of an MDO problem (adopted from Tosserams et al. 2010)

The coordination process consists essentially of two nested loops. The inner loop consists of the solution of each optimization subproblem using any appropriate optimization algorithm. The inner loop iterations refer to the number of analyses required by the optimization subproblems. The outer loop is responsible for collecting information from all subproblems to compute consistency constraint violations and update the linear and quadratic weights of the penalty functions. As mentioned earlier, MDO problems in real industrial environments, such as vehicle or aircraft design, require costly analyses (Kang et al. 2014b). Oftentimes, optimization studies are conducted with a limited budget of analyses due to time restrictions. Of course, the objective of such optimization studies is not to determine optimal designs, but to explore large design spaces relatively quickly, yet with reasonable accuracy; this is especially beneficial in the early design stages. If the disciplinary optimization subproblems require a large number of iterations to converge, the computational budget may allow only a few outer-loop iterations; this is, more often than not, insufficient to achieve acceptable consistency levels. For a sensible use of computational resources, it seems relevant to bound the number of iterations in each inner loop optimization.

This work introduces two integer parameters NI and NO to define the budget of maximum inner and outer iterations, respectively. A large value of NO may lead to higher MDO consistency. On the other hand, a large value of NI may yield better disciplinary designs. On the contrary, small values of NI may yield suboptimal disciplinary designs. This could lead to consistent but suboptimal MDO solutions. Balancing this trade-off may be possible by understanding the impact of the NI and NO parameter values. To do this, we first relate these two parameters to computational cost. The number of disciplinary optimization subproblems, the number of design variables in each subproblem and the computational cost of each analysis required by the optimization will have a major impact on the total MDO computational cost. For an MDO problem with j = 1,2,…,N sp disciplinary subproblems where each subproblem has n j design variables and t j analysis time, the total computational time of the process will be

$$ t_{total} = NO \times NI \times \sum\limits_{j=1}^{N_{sp}} \mathcal{O} (\sigma_{j}(n_{j})) t_{j}, $$
(1)

where σ j (n) depends on the optimization algorithm. For example, it can be the number of evaluations required to compute or approximate gradients and conduct a line search in a gradient-based algorithm or the number of orthogonal polling directions in a direct search algorithm. There are many values of NO and NI that can lead to similar total computational times. The main objective of this paper is to study the impact of the choice of NO and NI under a fixed budget B = NO×NI. In addition, we test 3 optimization algorithms to solve the individual subproblems. Finally, we also investigate the impact of two parameters used in the updating scheme of the linear and quadratic weights of the penalty functions.

The paper is organized as follows. The next section presents a revised notation of the non-hierarchical analytical target cascading (NHATC) formulation that aims at making it easier to understand, follow and implement. Section 3 presents numerical investigations conducted for four test problems. Section 4 provides summarizing remarks and suggestions for future work.

2 Non-hierarchical ATC formulation

A linking (also called coupling) variable y ij in an MDO problem is the output of disciplinary subproblem i and the input of disciplinary subproblem j. For such a variable, the subproblem that takes this variable as an input sets target values \(\mathbf {t}_{y_{ij}}\). The subproblem that yields this variable as an output reports response values \(\mathbf {r}_{y_{ij}}\). In the non-hierarchical ATC formulation, the coupling of a subproblem j to neighbor subproblems is defined by the set \(\mathcal {T}_{j}\) of neighbor subproblems to which subproblem j sends targets (and receives responses from them) and the set \(\mathcal {R}_{j}\) of neighbor subproblems to which subproblem j sends responses (and receives targets from them). At least one of the two sets must be non-empty for every subproblem, otherwise it is not linked to any other subproblem. In addition, a subproblem can both set targets to and receive targets from the same neighboring subproblem, i.e., a neighboring subproblem can belong to both sets. In such cases, the sent and received targets can only be related to different variables, i.e, y ij and y ji can obviously not be the same quantities.

In the example of Fig. 1, subproblem 1 is coupled to subproblems 2, 3 and 6: it receives inputs y 21 and y 31 from subproblems 2 and 3, respectively, and it sends outputs y 12 and y 16 to subproblems 2 and 6, respectively (i.e., \(\mathcal {T}_{1} = \{2, 3\}\) and \(\mathcal {R}_{1} = \{2, 6 \}\)). Therefore, it sets targets \(\mathbf {t}_{y_{21}}\) and \(\mathbf {t}_{y_{31}}\) to subproblems 2 and 3, respectively, and it reports responses \(\mathbf {r}_{y_{12}}\) and \(\mathbf {r}_{y_{16}}\) to subproblems 2 and 6, respectively. The following 4 target-response pairs must thus be coordinated in optimization subproblem 1: \(\mathbf {t}_{y_{21}}\) and \(\mathbf {r}_{y_{21}}\), \(\mathbf {t}_{y_{31}}\) and \(\mathbf {r}_{y_{31}}\), \(\mathbf {t}_{y_{12}}\) and \(\mathbf {r}_{y_{12}}\) and \(\mathbf {t}_{y_{16}}\) and \(\mathbf {r}_{y_{16}}\).

In addition to linking (also called coupling) variables, two subproblems may be linked by the existence of shared variables. However, shared variables are not the output of a disciplinary subproblem; they are common inputs to multiple disciplinary subproblems. The set \(\mathcal {S}_{j}\) is defined to include all neighbor subproblems that share variables with subproblem j (note that \(\mathcal {S}_{j}\) can be an empty set). When two subproblems i and j share variables, these are denoted by \(\mathbf {x}_{s_{ij}}\), while keeping i<j to avoid doublecounting. Then, a local copy is defined for each subproblem as \(\mathbf {x}_{s_{iji}}\) and \(\mathbf {x}_{s_{ijj}}\).

The formulation of (each) disciplinary optimization subproblem j, given updated information from all other subproblems (i.e., updated values for \(\mathbf {r}_{y_{ij}}, i \in \mathcal {T}_{j}\), \(\mathbf {t}_{y_{jk}}, k \in \mathcal {R}_{j}\), \(\mathbf {x}_{s_{jll}}, j<l \in \mathcal {S}_{j}\) and \(\mathbf {x}_{s_{ljl}}, j>l \in \mathcal {S}_{j}\)), is

$$ \begin{array}{rl} \min & f_{j}(\mathbf{x}_{j},\mathbf{t}_{y_{ij}},\mathbf{x}_{s_{jlj},j<l},\mathbf{x}_{s_{ljj},j>l}) + \\ & \hspace{0cm} \sum\limits_{i\in\mathcal{T}_{j}} \phi_{y_{ij}}(\mathbf{t}_{y_{ij}}-\mathbf{r}_{y_{ij}}) +\\ & \hspace{0cm} \sum\limits_{k\in\mathcal{R}_{j}} \phi_{y_{jk}}(\mathbf{t}_{y_{jk}}-\mathbf{r}_{y_{jk}}) +\\ & \hspace{0cm} \sum\limits_{l \in \mathcal{S}_{j}, j<l}\phi_{s_{jl}}(\mathbf{x}_{s_{jlj}}-\mathbf{x}_{s_{jll}}) +\\ & \hspace{0cm} \sum\limits_{l \in \mathcal{S}_{j}, j<l}\phi_{s_{lj}}(\mathbf{x}_{s_{ljl}}-\mathbf{x}_{s_{ljj}})\\ \text{wrt} & \mathbf{x}_{j},\mathbf{t}_{y_{ij}},\mathbf{x}_{s_{jlj},j<l},\mathbf{x}_{s_{ljj},j>l} \\ \text{st} & \mathbf{g}_{j}(\mathbf{x}_{j},\mathbf{t}_{y_{ij}},\mathbf{x}_{s_{jlj},j<l},\mathbf{x}_{s_{ljj},j>l})\leq \mathbf{0}\\ & \mathbf{h}_{j}(\mathbf{x}_{j},\mathbf{t}_{y_{ij}},\mathbf{x}_{s_{jlj},j<l},\mathbf{x}_{s_{ljj},j>l}) = \mathbf{0}\\ \text{where} & \mathbf{r}_{y_{jk}} = \mathbf{S}_{jk}\mathbf{a}_{j}(\mathbf{x}_{j},\mathbf{t}_{y_{ij}},\mathbf{x}_{s_{jlj},j<l},\mathbf{x}_{s_{ljj},j>l}), \\ & \hspace{0cm} k\in\mathcal{R}_{j}, i\in\mathcal{T}_{j}, l \in \mathcal{S}_{j}, \end{array} $$
(2)

where x j are local design variables. Here it is assumed that a single disciplinary analysis a j yields all responses of subproblem j; the binary selection matrix S jk is then used to select which responses are reported back to which neighbor subproblems. The two sums for the shared variables are required to keep the variable order correct in the penalty function of each subproblem. In the Appendix, we provide the formulations and analysis flow diagrams for all four examples considered in Section 3 to facilitate the interested reader’s comprehension of the formulation and its notation.

The quadratic penalty functions are defined as

$$ \phi_{(\cdot)_{ij}}(\mathbf{q})=\mathbf{v}_{(\cdot)_{ij}}^{T}\mathbf{q}+\|\mathbf{w}_{(\cdot)_{ij}} \circ \mathbf{q}\|_{2}^{2}, $$
(3)

where the subscript (⋅) is either y or s, q is the argument (steming from the consistency constraints q = 0), \(\mathbf {v}_{(\cdot )_{ij}}\) and \(\mathbf {w}_{(\cdot )_{ij}}\) are linear and quadratic weights, respectively, and ∘ denotes the Hadamard product (component-wise vector multiplication).

2.1 Coordination

The coordination process is depicted in Fig. 2. The outer loop collects the most recent target-response pair values from all subproblems, computes consistency constraint values (i.e., the difference between requested targets and reported responses), and updates the weights of all penalty functions using (4) and (5). The number of iterations of the outer loop is denoted NO; it is also referred to as the number of ATC iterations. The inner loop performs all disciplinary optimization subproblems, i.e., solves subproblem (2) associated to each discipline of the MDO problem. A budget of NI maximum iterations is allocated to each disciplinary optimization subproblem.

Fig. 2
figure 2

Non-hierarchical ATC coordination process

At an outer iteration k, the linear penalty weights v k are updated according to Tosserams et al. (2008) and Bertsekas (2003)

$$ {v_{i}^{k}}=v_{i}^{k-1} + 2~\left( w_{i}^{k-1}\right)^{2} q_{i}^{k-1}, $$
(4)

where q k−1 are the values of the consistency constraints after all disciplinary optimization subproblems have been solved using the penalty weights of iteration k−1.

The quadratic penalty weights w are then increased by a factor β if the reduction of the inconsistency is considered insufficient according to the following rule

$$ {w}_{i}^{k}=\left\{\begin{array}{ll} \;\;w_{i}^{k-1} & \text{if}~|q_{i}^{k-1}| \leq \gamma |q_{i}^{k-2}|\\ \beta w_{i}^{k-1} & \text{otherwise,} \end{array}\right. $$
(5)

where β>1 and 0≤γ≤1 are parameters held fixed during the entire ATC process. The values recommended in Tosserams et al. (2010) are β = 2.2 and γ = 0.4. The couple (β,γ) has an important impact on the efficiency of the coordination. This is investigated by means of numerical experiments in Section 3.6.

3 Numerical investigations

3.1 Test problems

Numerical experiments are performed for three analytical problems and one simulation-based problem (see Appendix A for detailed problem formulations). The main properties of these problems are listed in Table 1.

Table 1 Test problems

The first problem, called “bi-quadratic”, is a simple and smooth optimization problem with only one design variable that has been artificially decomposed to an MDO problem with three disciplinary subproblems.

The second problem is a geometric program with equality and inequality constraints. The equality constraints are used to decompose this problem to an MDO problem with three disciplinary subproblems (Kim 2001; Kim et al. 2003).

The third problem, is a simplified wing design problem with three disciplinary optimization subproblems. This problem has been formulated so that the design objectives (maximum take-off weight, fuel weight and wing weights) are similar to those of a 100-150 seat jetliner. The first disciplinary optimization subproblem aims at minimizing maximum take-off weight. The disciplinary optimization subproblems 2 and 3 minimize wing weight (which depends on the structural design of the wing) and fuel weight (which depends on the aerodynamic properties of the wing), respectively. The wing twist distribution is a design variable shared by both the structural and aerodynamic disciplinary subproblems. The planform is a design variable in disciplinary subproblem 2 (structures), but it is also required for the aerodynamic analysis in subsubproblem 3. The CST coefficients (Kulfan 2007) that enable fine-tuning the local shape of the wing are design variables in disciplinary subproblem 3 (aerodynamics).

The fourth problem, called “supersonic business jet”, is an MDO problem with 4 disciplinary optimization subproblems; it was used in Tosserams et al. (2010) to demonstrate non-hierarchical ATC formulations.

3.2 Optimization algorithm

Since MDO problems are typically simulation-based (and often blackbox) problems, rigorous derivative-free methods with convergence properties are desirable for the solution of the optimization subproblems. The optimization algorithm used in this work is based on the seminal paper on Mesh Adaptive Direct Search (Audet and Dennis 2006) and has been used in many engineering applications (Pourbagian et al. 2015; Gheribi et al. 2016; Aasi et al. 2013; Kang et al. 2014a; Spencer et al. 2013). The design variables of a given optimization problem are denoted by \(x\in {\mathbb {R}}^{n}\). Box constraints are provided such that \( \underline {x}_{i} \le x_{i} \le \bar {x}_{i} \;\forall i = 1...n\). A starting point x 0 which respects those bounds is provided. As depicted in Fig. 3, Δ p and \({\Delta }_{m} = \min \{{\Delta }_{p},{{\Delta }_{p}^{2}}\}\) are the poll and mesh sizes, respectively, used to generate poling directions. The mesh size quantifies the granularity between two possible polling directions, while the poll size defines the norm of these directions. The convergence of the MADS algorithm is guaranteed by the fact that the set of possible directions grows dense in the unit sphere as the mesh size decreases faster than the poll size (Audet and Dennis 2006; Clarke 1983). S is a diagonal matrix used for scaling. The initial poll size is set at \({\Delta }_{p}^{init}=1\). H is the set of polling directions and P is the polling set, i.e., the set of candidates evaluated at each iteration.

Fig. 3
figure 3

MADS poll and mesh sizes for a two-dimensional problem

We define as f min the objective function value of the best feasible point so far. The aggregate constraint \(h(x) = {\sum }_{j} \max \{c_{j}(x),0\}^{2}\) takes into account all nonlinear inequality and equality constraints c j (inequality constraints are satisfied when c j ≤0); x is feasible if and only if h(x)=0. We define as h min the aggregate constraint of the most feasible point found so far. If a point is not feasible, its objective is considered infinite. The detailed algorithm is presented in Fig. 4. Unlike more elaborate implementations of MADS (Talgorn et al. 2015; Audet et al. 2012; Le Digabel 2011), this version does not use optimistic direction, opportunistic evaluation, progressive barrier or surrogate models. This algorithm has the advantage of relying only on comparison between the objective values of two points. This is necessary when consistency is difficult to reach, as the penalty functions can reach very high values, which can cause other optimization algorithms to become inefficient or crash. Moreover, a tailored optimization algorithm such as this one provides an advantage in terms of computational time: for each problem and for each design evaluated, the objective, constraints and penalty functions are memorized separately. When several optimizations are run on the same problem, it is possible to keep the objective and constraints previously calculated, and thus reevaluate only the penalty function; this results in significant computation savings.

Fig. 4
figure 4

MADS algorithm

3.3 Test protocol

The inner-loop disciplinary optimizations are performed using a Matlab implementation of the MADS algorithm with orthogonal polling directions described in the previous subsection. The tolerance on the inequality constraints is set to 10−6. The maximum number of iterations allocated to each MADS call is 2×n j ×NI+1, where n j is the number of design variables in subproblem j. This allows the evaluation of the initial point, then exactly NI full iterations of the algorithm described in Section 3.2.

A set of 40 different initial guesses is used to conduct 40 optimization runs for each test problem. From these 40 runs, statistics are computed for two metrics: the inconsistency of the solution 𝜖 q and the discrepancy from the best known solution 𝜖 f :

$$\begin{array}{@{}rcl@{}} \epsilon_{q} & =& \underset{i}{\max} \left| \frac{q_{i}}{u_{i}-l_{i}} \right| \end{array} $$
(6)
$$\begin{array}{@{}rcl@{}} \epsilon_{f} & =& \frac{|f-f^{*}|}{|f^{*}|}, \end{array} $$
(7)

where u i and l i are appropriate quantities to normalize target-response pairs so that they are of the same order of magnitude. It is important to note that non-consistent solutions can be associated with better objective values. Therefore, it is necessary to compute the absolute value of the discrepancy from the best known solution and to always consider the inconsistency of the solution when assessing a set of parameter values.

3.4 Impact of NI and NO

In this section, the impact of the values of NI and NO on MDO solution quality are investigated. The couples (NI,NO) tested in this work are listed in Table 2. The product NI×NO is constant and equal to 4096.

Table 2 Tested values of NI and NO

The parameters are set to β = 2.2 and γ = 0.4 as suggested in Tosserams et al. (2010). In all numerical experiments, the initial penalty weight values are v 0=0 and w 0=1.

Figure 5 shows the evolution of 𝜖 q and 𝜖 f during the optimization. Each point of each curve represents the state of the optimization at the end of an outer-loop iteration. The x-axis reports cumulative number of iterations; the distance between two points is thus equal to NI. As the penalty weights are updated at each outer-loop iteration, it can be observed that larger NO values yield higher MDO solution consistency. However, except for the very simple bi-quadratic problem, large NO values lead to poor MDO solution quality as measured by discrepancy from best known solution. This is caused by the small iteration budget allocated in each inner-loop optimization: MDO is driven by consistency. Accordingly, we observe that small NI values lead to poor MDO solution quality (especially for NI = 8 where the 𝜖 f curves are nearly flat). On the contrary, it appears that for a small number of outer-loop iterations, the penalty function terms are dominated by local design objective terms. This also leads to poor consistency and thus to a large discrepancy from best known solutions. All numerical results are provided in Appendix B.

Fig. 5
figure 5

Evolution of inconsistency and discrepancy from best known solution

Figure 6 depicts the statistical distributions of MDO solution consistency and quality at the end of the ATC process. Each subfigure represents the distribution of 𝜖 q and 𝜖 f over the 40 runs for a given couple of values (NI,NO). The thick horizontal line indicates the median value. The top and bottom of the box indicate the 25th and 75th percentiles. The short horizontal lines indicate the 10th an 90th percentiles. To allow the display in logarithmic scale, values of 𝜖 q smaller than 1e-20 are displayed as being 1e-20. For the three first problems, we observe a clear trend: moderate values of NI and NO lead to good consistency. Indeed, these parameters allow both to have enough outer loop iterations to increase the penalty parameters v and w, and to have enough inner loop iterations to efficiently minimize the penalty function. For the most challenging supersonic business jet problem, which has a very narrow feasible domain, moderate to low values of NI lead to a slightly better consistency. We observe that for NI = 512, inconsistency is high. As a consequence, 𝜖 f is also high as the ATC process converged toward poorly-consistent MDO solution, which can result in large deviations of f relative to f . This effect, along with a very narrow range of values, can be observed for NI = 512 for all problems. As opposed to the first two problems (which are contrived MDO problems), for the two “natural” MDO problems (the simplified wing design and supersonic business jet), we observe a dramatic reduction of the range of 𝜖 f values for NI = 512. This can be explained by the fact that a large value of NI leads to a very good convergence of the inner-loop optimization (hence the narrow range of values), but the small value of NO leads to a poor consistency and consequently a poor value of 𝜖 f . This leads to the conclusion that too much effort has been spent in each inner-loop optimization, which is detrimental to both consistency and quality of the MDO solution.

Fig. 6
figure 6

Inconsistency and discrepancy from best known solution as a function of NI and NO

For the first three problems, it is clear that consistency is best when there is a balance between NI and NO. For the supersonic business jet, this balance seem to be of less importance, yet high NI values seem to be detrimental. In most cases, and in particular for the two last problems, the value of NI does not seem to have a great impact on the discrepancy 𝜖 f . However, a design is only valid if it has a good consistency. Moreover, the measures 𝜖 q and 𝜖 f are linked. In particular, in Fig. 5, we often observe a bounce of 𝜖 f before it converges. This bounce is caused by the absolute sign in the definition of 𝜖 f in (7). With a very inconsistent design, the objective f is generally smaller than f . Then the objective increases as the consistency improves, which leads to a diminution of 𝜖 f , until f crosses f , then 𝜖 f starts to grow again.

These observations lead us to choose the values NI = NO = 64 for the investigation of the parameters β and γ in Section 3.6. Finally, it is important to remember that all the presented results are associated with a fixed computational budget. The fact that the most challenging supersonic business jet problem did not achieve as high consistency as the other three within the fixed coimputational budget merely manifests the need for more coordination iterations. Results obtained with unlimited budgets were reported in Tosserams et al. (2010). Nevertheless, in Section 3.7 we report high-consistency results for the supersonic business jet problem obtained by means of an increased computational budget.

3.5 Using modified MADS and gradient-based interior-point algorithms

In this section, the results obtained using the MADS algorithm are compared to those obtained using two other optimization algorithms. First, we use the algorithm MADSΔ which is a variation of MADS with a novel feature that defines the initial poll size of each problem optimization. This feature is motivated by the the loss of efficiency caused by the sequential optimization (i.e., the interruption of the optimization to update the penalty functions). In particular, the poll size in the MADS algorithm decreases as the optimization unfolds. However, when the optimization is interrupted and restarted (by taking the last design returned as the new starting point) the poll size is re-initialized at 1. To have a perfect continuity between two sequential optimizations (if the two optimization problems were perfectly identical), the poll size would need to be initialized with the final value of the previous algorithm. To improve the convergence, we propose the following definition of the initial poll size:

$$ {\Delta}_{p}^{init} = \left\{ \begin{array}{c l} 1 & \text{if}\; {\Delta}_{p}^{max}\; \text{is undefined or equal to 0}, \\ 2{\Delta}_{p}^{max} & \text{otherwise}, \end{array}\right. $$
(8)

where \({\Delta }_{p}^{max}\) is the largest poll size that led to a success in the previous optimization of the same problem. In other words, the value \({\Delta }_{p}^{max}\) that is returned at the end of an optimization is an estimate of (is of the the same order of magnitude than) the width of the attraction basin at the beginning of this optimization. The factor 2 aims at improving global exploration.

Then, we compare MADS and MADSΔ to the interior-point algorithm implemented in Matlab’s fmincon function. For each fmincon optimization, the evaluation budget is the same as that defined for MADS (2×n×NI). The performance of these three algorithms is displayed in Fig. 7 in a fashion similar to that of Fig. 6. For each problem, optimization method and value of NI, the 10th, 25th, 75th, 90th percentiles and the median value are represented for the inconsistency 𝜖 q and for the discrepancy from best known solution 𝜖 f .

Fig. 7
figure 7

Comparison of 3 methods for the optimization of the problems: MADS (white, left of each column), MADSΔ (light gray, center) and fmincon (dark gray, right)

For small values of NI the MADSΔ algorithm leads to very good values of consistency for the three first problems. However, for large values of NI, the advantage of having a good initialization of the poll size is muted by the large number of iterations available for the optimization and MADSΔ is outperformed by MADS. Moreover, MADSΔ is predominantly a local method, which leads to poor values of 𝜖 f . It seems that MADSΔ performance could be improved by gradually changing \({\Delta }_{p}^{init}\) from 1 to \(2{\Delta }_{p}^{max}\) as the ATC process unfolds. In addition to that, the very local nature of MADSΔ may keep it from reaching consistency. For the bi-quadratic problem, the objective is convex so MADSΔ always leads to a very good consistency for NI≤32. But for the geometric programming and simplified wing design problems, the non-convexity in the objective causes an unexpected behavior of MADSΔ: it will reach 1e-20 in 50% of the runs but will be stuck in a local inconsistent minimum in the other 50 %.

For the challenging supersonic business jet problem, MADSΔ does not improve consistency with the fixed computational budget; fmincon tends to yield better consistency than MADS but worse objective values. Indeed, MADS is more able to explore the nonsmooth design space, but when the consistency is not reached, the objective function becomes mainly quadratic and fmincon is then more efficient at improving consistency. For MDO applications with non-smooth subproblems, MADS seems to be the most efficient method. However, it must be noted than in general, the optimization algorithm can be different for each subproblem depending on its characteristics.

3.6 Impact of β and γ

The following values have been used for the parameters β and γ of the updating scheme of the quadratic penalty weight: γ∈{0.0;0.2;0.4;0.6;0.8;1.0} and β∈{1.4;1.8;2.2;2.6}. Figure 8 depicts 𝜖 q and 𝜖 f values for each test problem and each possible couple (β,γ). Equation (5)implies that larger β values and/or smaller γ values yield high consistency. Thus, 𝜖 q values should be smaller for the bars in the front of the figure and higher for those in the back. The results shown in Fig. 8 seem to be aligned with this trend, and confirm the recommendation of (Tosserams et al. 2010).

Fig. 8
figure 8

Inconsistency and discrepancy from best known solution as a function of β and γ

3.7 Additional results for the supersonic business jet problem

As mentioned in Section 3.4, the allocated computational budget is not sufficient to achieve high consistency for the supersonic business jet problem. Therefore, we increase it to be four (4) times higher, namely NI×NO = 16384. The values of NI and NO tested are listed in Table 3. We used the values β = 1.4 and γ = 0.4 which seemed to yield high consistency in previous runs (see Fig. 8). The results are presemted in Fig. 9.

Table 3 Values of NI and NO for the supersonic business jet with increased computational budget
Fig. 9
figure 9

Inconsistency and discrepancy for the supersonic business jet problem with an increased computational budget

4 Concluding remarks

The work presented in this paper consists of three parts. We first investigated the effect of the number of coordination iterations (NO) and optimization iterations (NI) on the quality of the solution of MDO problems for fixed computational budgets. Solution quality was quantified by consistency and closeness to best known solution. Two contrived and two “natural” MDO test problems were used. It seems that a balanced number of moderate NO and NI values yields best results for most comsidered problems. Extreme values of NI and NO lead to inconsistent or suboptimal MDO designs. However, NO must be large enough so that, as the coordination process progresses, the gradient of the penalty functions is adequately larger than the gradient of the local design objectives.

We then considered both rigorous derivative-free and gradient-based algorithms to perform the subproblem optimizations. The MADS and MADSΔ derivative-free algorithms seem to yield higher-consistency solutions, but with greater variance, than the Matlab interior-point gradient-based algorithm, with the exception of the supersonic business jet problem. Conversely, the interior-point method seems to yield higher-quality solutions, again with the exception of the supersonic business jet problem. A method to initialize the poll size was proposed that leads to good consistency but may be too local to yield the best objective values.

Finally, numerical experiments were conducted to determine guideline values for the parameters β and γ of the updating scheme related to the quadratic weights of the penalty functions. It seems that values from the following ranges yield higher-consistency solutions for most considered problems: β∈[2.2,2.6] and γ∈[0.0,0.4]. For more difficult problems like the supersonic business jet problem, higher computational budgets are required; in addition, smaller values of β seem to yield better results.

In future work, the initialization of the quadratic weight (w 0) should be investigated; we hypothesize that a gradient norm of the penalty function that is slightly lower than that of the local objective in the beginning of the optimization can be beneficial to the performance of the coordination algorithm. This would allow to achieve consistent MDO designs within a smaller number of outer loop iterations, without adverse effect on optimality. Moreover, the initialization of the poll size can be improved to remedy the loss of efficiency due to sequential optimization. Finally, it should be investigated whether a dynamic management of NI as the coordination process progresses can improve both consistency and optimality of the final MDO design.