1 Introduction

The field of metaheuristics has emerged as a promising research area in which those involved focus on solving various optimization problems by using the simulation of a variety of natural phenomena. Genetic algorithm (GA) [1], particle swarm optimization (PSO) [2], and differential evolution (DE) [3] are a few examples of some early metaheuristic algorithms that have been proposed throughout the past decades. Metaheuristic algorithms have now been used in many research areas, including those of computer science [4], forecasting [5, 6], medicine [7], economics [8], and engineering [9, 10]. The success of those metaheuristic algorithms in solving various optimization problems across a wide range of fields has resulted in proposals for promising algorithms like artificial bee colony (ABC) [11], multi-verse optimizer (MVO) [12], symbiotic organisms search (SOS) [13], and Jaya algorithm [14]. One of the advantages of metaheuristic algorithms is that they do not need to deeply adapt to each problem in order to solve it, which is contrary to problem-specific heuristics [15]. According to the “No-Free-Lunch” theorem [16], any pair of search algorithms has similar average performance with regard to all possible optimization problems. However, this does not exclude the possibility that certain algorithms will obtain better results for some objective functions. Thus, the need to develop high-performance metaheuristic algorithms increases continually.

Inspired by the idea of the classroom teaching–learning process, Rao et al. developed a metaheuristic algorithm entitled the teaching–learning-based optimization (TLBO) algorithm to solve various constrained [17] and unconstrained benchmark test functions [18]. TLBO mimics the natural phenomena of the classroom. For the TLBO algorithm, the term “student” is used to represent a candidate solution to the corresponding optimization problem. Meanwhile, the term “teacher” deemed to be the best learner in the classroom, and it is used to represent the current best solution. All students in the classroom individually perform unique search patterns over solution space through the teacher phase and the learner phase. The teacher phase is characterized by the transmission of knowledge from a teacher to students. The learner phase, on the other hand, is characterized by the transmission of knowledge from mutual interactions between students. Metaheuristic algorithms in general require extra algorithm-specific control parameters. Thus, the advantage the TLBO has over majority of metaheuristic algorithms is that it only requires control parameters such as maximum number of generations and population size.

Since it was initially proposed, there have been numerous applications of the TLBO in solving different complex problems. Some are utilized in data clustering [19], mechanical engineering [17], structural engineering [20], parameter optimization [21], and economic dispatch [22]. Several discussions related to TLBO have also appeared in the literature surrounding this topic [23, 24]. As it is a new and promising optimization algorithm, researchers worldwide have attempted to enhance the performance of TLBO. Rao and Patel [25] added the elitism concept, where the elite solutions replace the worst solutions. To make TLBO faster and more robust, Satapathy and Naik [26] introduced a modified TLBO (mTLBO), which altered the learner phase formula by adopting extra training through tutorials. Rao and Patel [27] presented several new elements in the improved TLBO (I-TLBO) such as adaptive teaching factors, number of teachers, tutorial training, and self-motivated learning to increase the searching ability of TLBO. Previous research findings have successfully improved the capability of the basic TLBO, and thus, developing better metaheuristic algorithms remains a top priority as the amount of hard and complex problems have continued to augment.

The fact that the TLBO algorithm is relatively new yet demonstrates superior performance over other metaheuristics algorithms in solving various benchmarks and engineering problems, has drawn our attention and encouraged us to further investigate and enhance the capability of the TLBO algorithm. In this paper, a new metaheuristic algorithm called fuzzy adaptive teaching–learning-based optimization (FATLBO) is introduced to improve the performance of its predecessor. Three new modifications are implemented: a status monitor, fuzzy adaptive teaching–learning strategies (FATLS), and a remedial operator. These modifications aim to increase the effectiveness and efficiency of TLBO. In the status monitor, a new procedure to track the productivity of each student in both the teacher and learner phases is developed. FATLS uses a fuzzy logic controller to control the operation of teacher and learner phases, setting a much larger weight on the most productive phase and thus increasing the convergence speed. A remedial operator is deployed to activate the restart mechanism for the entire classroom if there is no significant improvement over a certain period of time. We compare FATLBO with a number of algorithms, including the basic TLBO, ITLBO, and other well-known metaheuristic algorithms across a range of benchmark functions. Experimental results confirm the excellent performance of proposed FATLBO algorithm in comparison with the other metaheuristic algorithms.

The remaining sections are organized as follows. In Sect. 2, we describe the regular TLBO algorithm. In Sect. 3, we briefly review previous works related to parameter adaptation techniques. In Sect. 4, we describe the proposed algorithm in further detail. In Sect. 5, we present the experimental results. Finally, Sect. 6 consists of our conclusions.

2 The teaching–learning-based optimization (TLBO) algorithm

Rao et al. [17, 18] introduced TLBO as a population-based metaheuristic algorithm that simulates traditional classroom learning. First, a random population consisting of students is generated based on population size and the number of design variables. A student, X i , is represented by one solution vector to a specific objective function with a corresponding fitness value, f(X i ). The searching procedure is performed throughout two main phases of the learning process: the teacher phase and the learner phase. The efficiency of knowledge transfer is determined by the appropriate combination of these two phases. In the teacher phase, a teacher is required to deliver knowledge directly to students. After evaluating the fitness value of each student, the teacher (X teacher) role is assigned to the smartest student with the lowest fitness value in that iteration. The other students are then improved by shifting the average knowledge of all students (X mean) toward the X teacher as shown in Eq. (1):

$$X_{{i{\text{ new}}}} = X_{i} + {\text{rand}} \times \left( {X_{\text{teacher}} - T_{F} \times X_{\text{mean}} } \right),$$
(1)

where rand is a uniform random number between 0 and 1 and T F is a teaching factor with a random value of either 1 or 2.

In the learner phase, a student (X i ) engages with other students and benefits from the exchange, which is selected randomly (X j ). The student X i is first compared with student X j . If student X j is smarter than X i , then X i is shifted toward X j as shown in Eq. (2). Otherwise, it is shifted away from X j as shown in Eq. (3):

$$X_{{i{\text{ new}}}} = X_{i} + {\text{rand}} \times \left( {X_{j} - X_{i} } \right),$$
(2)
$$X_{{i{\text{ new}}}} = X_{i} + {\text{rand}} \times \left( {X_{i} - X_{j} } \right).$$
(3)

For both the teacher and learner phases, the fitness value of each student is allowed to update only when the new solution (X i new) outperforms the former one (X i ). Subsequently, the TLBO attempts to enhance some students by modifying them throughout the teacher and learner phases. The algorithm is run until the maximum number of iterations is reached. The whole main procedure of the TLBO algorithm is shown in Fig. 1.

Fig. 1
figure 1

TLBO flowchart

3 Previous works related to parameter adaptation techniques

It is worth noting that although TLBO is described in the literature as an algorithm-specific parameter-less optimization method, common control parameters (such as population size and maximum number of iterations) still need to be determined. Nevertheless, unlike most metaheuristic algorithms, TLBO does not have any algorithm-specific control parameter. Hence, no additional effort is required to set additional parameters for the algorithm to work. This situation offers a huge benefit to TLBO since choosing the proper control parameters requires some expertise. The inappropriate choice of control parameters may contribute to worse performance.

Despite these above-mentioned advantages, the lack of additional algorithm-specific parameters also entails some drawbacks. The additional algorithm-specific parameters help metaheuristic algorithms generate some unique searching patterns to solve various optimization problems comprising a variety of fitness landscapes [28]. In GA, after the selection process, the solutions undergo crossover and mutation operators with certain probabilities. The crossover and mutation operators occur only within some probabilities, called the crossover rate (CR) and the mutation rate (MR), respectively. When the solutions are not subjected to crossover or mutation, they remain unmodified. CR and MR are categorized as algorithm-specific control parameters in GA. The choice of CR and MR is known to significantly affect the behavior and performance of GA [29]. Similarly to GA, TLBO has two operators that modify the current solutions, known as teacher phase and learner phase. However, unlike GA, these two operators are not subjected to probabilities. The solutions will undergo teacher and learner phases throughout the iterations. With respect to CR and MR, adding algorithm-specific parameters that control the probabilities of both the teacher and learner phases may increase the performance of TLBO.

Many researchers have developed parameter adaptation techniques for finding the optimal parameter setting instead of setting fixed values to the control parameters. For example, Qin et al. [30] proposed a self-adaptive DE algorithm (SaDE) that employs mutation strategies and the respective control parameter is self-adapted based on their previous experiences of generating promising solutions. Yang and Hsieh [31] developed auto-tuning boundary-approaching PSO (AB-PSO) algorithm, which has the feature of adaptive auto-tuning parameters. The adjustment is conducted after a certain number of iterations, allowing some time for the control parameters to make an impact.

Meanwhile, some researchers have experimented with using fuzzy concepts to improve the performance of different types of optimization algorithms. Shi and Eberhart [32] proposed fuzzy adaptive PSO to improve the performance of the PSO by dynamically adjusting the inertia weight parameter by integrating a fuzzy system with fitness feedback of each particle. Melin et al. [33] improved the swarm diversity and convergence capability of PSO using fuzzy parameter adaptation for designing the fuzzy classifier. Several schemes of Mamdani’s fuzzy inference systems were developed to adjust the cognitive factor (c1) and social factor (c2) parameters using the information of iteration, swarm diversity, and swarm convergence. Using fuzzy logic controllers whose inputs incorporate the relative function values, Liu and Lampinen [34] proposed fuzzy adaptive DE (FADE) to make some parameter adjustments for the mutation and crossover operations. The obtained results show that the FADE algorithm has a better performance than a regular DE in regard to higher-dimensional problems.

In summary, the setting of algorithm-specific parameters with the appropriate adaptive strategy may affect the optimization performance of metaheuristic algorithms significantly. Introducing some new algorithm-specific control parameters and implementing the right strategy for its parameter adaptation may improve the performance of TLBO.

4 Fuzzy adaptive teaching–learning-based optimization (FATLBO)

In the real classroom, teaching methods are paramount for the successful transmission of teachers’ knowledge to students. Certain materials are adapted to the teacher-centered approach, whereas the others are adequate for the so-called student-centered approach. The term student-centered approach refers to the teaching style in which students are involved in discussions and learning through cooperation. It is important to underline that inadequate teaching methods are likely to have an adverse effect on the learning process.

In the fundamental form of the TLBO algorithm, students are required to experience both teacher and learner phases. Accordingly, the TLBO needs two function evaluations to calculate the fitness value of one candidate solution during one iteration. This paper presents the implementation of the teaching methods used in the real classroom through the usage of modifications of the fundamental TLBO. Section 4.1 will elaborate on the status monitor we developed. This monitor records the progress of students related to both phases. It has the capacity to measure the productivity in both phases, as well as the improvement of the knowledge of students in a given time frame of an iteration. Section 4.2 presents the new strategy called the fuzzy adaptive teaching–learning system (FATLS), to choose appropriate teaching methods in order to optimize the efficiency and the search speed of the TLBO algorithm. We allow the students to skip the teacher or learner phase by introducing two new parameters, called the teaching rate (TR) and the learning rate (LR). TR and LR present the probabilities that a student will enter the teacher and learner phases. If TR is set to 0.7, then every student will have a 70 % probability of moving to the teacher phase and 30 % probability of skipping the teacher phase. Fuzzy logic is further employed to self-adjust the TR and LR based on information collected from the status monitor. Finally, in Sect. 4.3, we introduce a remedial operator to ensure students are not trapped in local optima.

4.1 Status monitor

Two new variables are introduced to measure the productivity of students during the teacher and learner phases: the success rate of teacher phase (SRTP) and the success rate of learner phase (SRLP). The values of SRTP and SRLP vary between 0 and 1, where a higher value means the more success in improving student knowledge.

Table 1 explains the impact of the SRTP and the SRLP calculation during one iteration. As can be seen from Table 1, the value of the SRTP is 0.4, whereas the value of the SRLP is 0.6. These results show that entering the teacher phase is going to enhance the fitness for about 40 % of students and entering the learner phase is going to enhance the fitness for about 60 % of students. The advantage of this monitor is that it follows the progress of students in all iterations. FATLS uses the SRTP and the SRLP information obtained by the status monitor to adapt the fittest TR and LR for each student in the following step.

Table 1 Example of success rate calculation

4.2 Fuzzy adaptive teaching–learning strategies (FATLS)

The fuzzy adaptive system has been developed to boost productivity during the learning process. The SRTP and SRLP information gathered by the status monitor is further used as input variables. In addition, regarding output variables, we proposed a bar controller to adapt the LR and the TR. Figure 2 represents an illustration of this bar controller. The central bar is actually the navigation bar. When the bar goes toward the left, the TR is in decline. It practically means that the FATLS will focus on the learner phase. On the contrary, when the bar goes toward the right, the LR will be in decline. Bar movement is used as the output variable. Then, we adjust the bar movement to tune the TR and LR.

Fig. 2
figure 2

Fuzzy bar controller

The FATLS consists of three principal components: fuzzification, inferencing process, and defuzzification.

4.2.1 Fuzzification

Figure 3 shows the conversion of all input variables from the preceding step into the membership grade. Left-triangle, triangle, and right-triangle functions are used as the membership functions for all input variables. Figure 4 shows five triangular membership functions used for the output variable, namely one left-triangle, three triangles, and one right-triangle.

Fig. 3
figure 3

Membership function of the SRTP and SRLP

Fig. 4
figure 4

Membership function of the bar movement

4.2.2 Inferencing process

The Mamdani-type fuzzy rules are employed to formulate the conditional statements that consist of the fuzzy rule base. The Mamdani’s fuzzy inference system is also applied to map all the input variables to the target output. As previously mentioned, the input variables are success rate of teacher phase (SRTP) and success rate of learner phase (SRLP), while the output variable is bar movement. The bar movement is critical to the tuning of the teaching rate (TR) and learning rate (LR) parameters. The proposed mapping process by the fuzzy inference system is illustrated in Fig. 5. Meanwhile, the following nine fuzzy rules aim to develop the best parameter adaptation scheme for the TR and LR parameters:

Fig. 5
figure 5

The mapping process by Mamdani’s fuzzy inference system

  1. 1.

    If (SRTP is LOW) and (SRLP is LOW), then (bar movement is NEUTRAL).

  2. 2.

    If (SRTP is LOW) and (SRLP is MEDIUM), then (bar movement is LEFT).

  3. 3.

    If (SRTP is LOW) and (SRLP is HIGH), then (bar movement is FAR-LEFT).

  4. 4.

    If (SRTP is MEDIUM) and (SRLP is LOW), then (bar movement is RIGHT).

  5. 5.

    If (SRTP is MEDIUM) and (SRLP is MEDIUM), then (bar movement is NEUTRAL).

  6. 6.

    If (SRTP is MEDIUM) and (SRLP is HIGH), then (bar movement is LEFT).

  7. 7.

    If (SRTP is HIGH) and (SRLP is LOW), then (bar movement is FAR-RIGHT).

  8. 8.

    If (SRTP is HIGH) and (SRLP is MEDIUM), then (bar movement is RIGHT).

  9. 9.

    If (SRTP is HIGH) and (SRLP is HIGH), then (bar movement is NEUTRAL).

All the fuzzy rules are designed to choose the teaching method that will have the most impact on the learning process. The bar controller plays a key role in controlling the values of TR and LR. If the bar moves to the left, then it lowers the TR and maintains the high rate of LR. This forces the students to attend the learner phase more frequently than the teacher phase, thereby mimicking a student-centered approach. Conversely, if the bar moves to the right, then it lowers the LR and maintains the high rate of TR. This forces students to attend the teacher phase more frequently than the learner phase, thereby creating a teaching-centered approach. In the inferencing process, the bar movement will adapt through different combinations of SRTP and SRLP, which were previously calculated from the status monitor. When SRTP is high and SRLP is low, it indicates that students’ knowledge is improving mainly due to the effect of the teacher phase rather than that of the learner phase. Therefore, it is reasonable for FATLBO to switch the current teaching–learning strategy into a more teacher-centered approach by moving the bar controller to the right (decreasing LR while maintaining TR) and vice versa.

4.2.3 Defuzzification

The next part of the process is the opposite of fuzzification. After the process of inferencing is complete, defuzzification is begun. Center-of-sums is the method we chose for defuzzification. Fuzzy strategies have an output, named bar movement, and based on that, the LR and TR of the following iterations can be adjusted. As an example, if there is a bar movement result of −0.35, the bar is going to travel 3.5 steps toward the left, which means that the TR and LR will be adjusted to 0.65 and 1.0, respectively, as shown in Fig. 6.

Fig. 6
figure 6

Fuzzy bar controller after FATLS applied

4.3 Remedial operator

There are times in class when whole groups of students show no improvement in the process of learning. This could be due to several factors, such as bad environment, difficult topic, or teaching style. Faced with this issue, the teacher could evaluate it and as a result add remediation or perturbation to the classroom based on the current circumstances. Remediation such as this may include downgrading the content and/or volume of the lesson, the application of personal tutoring, or a change in the style of teaching.

On the occurrence of premature convergence, the remedial operator adjusts the environment of the classroom by a population expansion utilizing the random step based on the Gaussian law, with a center point of the teacher. In this paper, we assume premature convergence if fitness value does not improve after a long period or if the whole population makes a convergence in a single solution.

The formula for the remedial operator is shown in Eq. (4).

$$X_{{i{\text{ new}}}} = X_{\text{teacher}} + N(0,{\text{range}}/2),$$
(4)

where range is equal to upper and lower bounds.

To conclude, the entire procedure of FATLBO is shown in Fig. 7.

Fig. 7
figure 7

FATLBO flowchart

5 Experimental results

We compare the FATLBO performance with other metaheuristic algorithms from the literature, including TLBO, ABC, PSO, DE, and their variants. This section is typically divided into four subsections that explain the four different experiment sets, including the details regarding benchmark functions, the setup of the experiment, and the results.

5.1 Experiment 1

In this experiment, we apply FATLBO to eight different benchmark functions that were originally solved by Liang et al. [35]. Table 2 gives the details of the benchmark functions. Previously, all functions were tested with the maximum of 30,000 function evaluations. To maintain consistency in the comparison, the FATLBO algorithm is also tested with the same maximum number of function evaluations. For the population size, FATLBO uses 10 individuals for all functions except Function 2, which uses 25 individuals.

Table 2 Details of benchmark functions in Experiment 1 (D: dimensions)

Each benchmark function is tested 30 times with the FATLBO algorithm, and results are obtained in the form of mean value and standard deviation. FATLBO is compared with TLBO variants [27], ABC variants [36], and PSO variants [35]. With the exception of FATLBO, the rest of the results are obtained from those previous works. The comparative results of FATLBO and other algorithms are shown in Table 3.

Table 3 Comparative results of TLBO and other algorithms in Experiment 1

The FATLBO algorithm identified better solutions compared with other algorithms in 6 functions: Sphere, Rosenbrock, Griewank, Weierstrass, Rastrigin, and NCRastrigin functions. Meanwhile, I-TLBO, TLBO, CLPSO, and modified ABC have only found 5, 5, 4, and 1 solution(s) better than others, respectively. FATLBO is able to achieve the global optimum in the Sphere, Griewank, Weierstrass, Rastrigin, and NCRastrigin functions. Furthermore, the FATLBO algorithm performs better than the rest of the considered algorithms for the Rosenbrock function. Although FATLBO was not able to find a solution better than other algorithms in the Ackley and Schwefel functions, FATLBO can consistently produce a strong performance and is still better than more than half of its competitors.

Although the results presented in Table 3 provide first insight into the performance of the algorithms, carrying out a statistical hypothesis test is still required for a better comparison. For this purpose, an unpaired t test is performed with a statistical significance value α = 0.05. Table 4 shows the statistical results of FATLBO compared with those of other algorithms. The last column of Table 4 presents the total count of the three statistically significant conclusions of performance comparisons between FATLBO and other algorithms from the t test (a/b/c). According to the table, FATLBO outperforms the rest of the algorithms, including the PSO and variants, the original TLBO, and the improved version of TLBO (I-TLBO). In comparison with FATLBO, the second best algorithm is I-TLBO, where FATLBO has a significantly better performance in two functions, a statistically similar performance in five functions, and a significantly worse performance in only one function.

Table 4 Statistical comparison of FATLBO with other optimization algorithms using an unpaired t test in Experiment 1

5.2 Experiment 2

In this experiment, we apply FATLBO to eight different benchmark functions that Mathur et al. [37] previously solved. The benchmark function set is solved later by Pham et al. [38], Ahrari and Atai [39], and Rao and Patel [27]. The results obtained using the FATLBO algorithm are compared with the previous works of TLBO variants along with other well-known optimization algorithms. The details of the benchmark functions are shown in Table 5.

Table 5 Details of benchmark functions in Experiment 2 (D: dimensions)

We maintain consistency in the comparison between previous works by adhering to the same rules. In previous studies, the optimization process stops when the difference between the fitness value and the global solution is lower than 0.1 % or less than 0.001, whichever is smaller. FATLBO uses a population size of 40. Each benchmark function is tested 100 times, and Table 6 shows the comparative results in the form of mean of function evaluations and success percentages. The results of the other competing algorithms are obtained from previous works.

Table 6 Comparative results of TLBO and other algorithms in Experiment 2

Table 5 demonstrates that the FATLBO algorithm is able to produce the fewest number of function evaluations in 7 out of 8 problems. Moreover, FATLBO requires approximately 29 and 52 % fewer function evaluations in the De Jong function (F1) than I-TLBO and TLBO, respectively; 17 and 32 % in the Branin function (F3); 30 and 61 % in the Martin and Gaddy function (F4); 13 and 32 % in the Rosenbrock1 function (F5); 8 and 32 % in the Rosenbrock2 function (F6); 61 and 68 % in the Rosenbrock3 function (F7); and 14 and 23 % in the Hyper Sphere function (F8). The only function where FATLBO failed to achieve the fewest number of function evaluations is the Goldstein and Price function (F2). However, FATLBO still produces the second fewest number of function evaluations after I-TLBO.

5.3 Experiment 3

In this experiment, the FATLBO algorithm is compared with TLBO and ABC variants in a small scale to a large scale by considering the dimensions 20, 30, and 50 for all the benchmark functions. In this part of the work, TLBO and I-TLBO are tested on 13 unconstrained benchmark functions, an approach that was first attempted by Li et al. [40] and later by Rao and Patel [27].

The complete formulation of these benchmark functions is given in Table 7. All functions have the global minimum 0 on every dimension except the Schwefel function (F8). The first seven functions (F1–F7) are categorized as unimodal, and the remainder (F8–F13) is multimodal.

Table 7 Details of benchmark functions in Experiments 3 and 4 (D: dimensions)

Each benchmark function is tested 30 times with the FATLBO algorithm, and the results are obtained in the form of mean solution and standard deviation. The maximum number of function evaluations is 40,000. The population size of FATLBO is 10 for all functions except F5, F12, and F13, which use a population of 40.

Table 8 shows the comparative results of FATLBO, I-TLBO, TLBO, GABC, ABC, and I-ABC algorithms for 13 functions with the maximum of 40,000 function evaluations. Except for FATLBO, the rest of the results are obtained from the previous works. The experiment revealed that the FATLBO algorithm outperformed the rest of the considered algorithms. FATLBO achieved the highest score in finding solutions more effectively than others (26) and surpassed the newly developed I-TLBO (19), I-ABC (17), and the original TLBO (9). It can be noted that FATLBO produces a better solution precision, especially in solving the Step function as well as the Quartic function. It is also possible to see that the success of FATLBO in solving the benchmark functions is not over-sensitive to the function dimension or the function characteristic. FATLBO performs well in various dimensions and unimodal or multimodal functions.

Table 8 Comparative results of TLBO and others algorithm in Experiment 3

Table 9 shows the statistical results of FATLBO compared with those of other algorithms using an unpaired t test with a statistical significance value α = 0.05. The last row of Table 9 presents the total count of the three statistically significant conclusions of performance comparison between FATLBO and the other algorithms from the t test (a/b/c). According to the table, FATLBO outperforms the other algorithms, including the ABC and variants, the original TLBO, and the improved version of TLBO (I-TLBO). In comparison with FATLBO, the second best algorithm is I-TLBO where FATLBO has significantly better performance in 14 functions, statistically similar performance in 18 functions, and significantly worse performance in 7 functions.

Table 9 Statistical comparison of FATLBO with other optimization algorithms using an unpaired t test in Experiment 3

5.4 Experiment 4

This experiment is intended to identify the performance of FATLBO in solving complex and high-dimension problems. The same thirteen benchmark functions from Experiment 3 are tested using the FATLBO and TLBO algorithms, each with 500 and 1000 dimensions, which Shi et al. [41] originally solved. For 500-D functions, 2.5E6 maximum function evaluations are used. For 1000-D functions, 5E6 maximum function evaluations are used. To ensure that the comparison is consistent, the same maximum number of function evaluations was used to test both TLBO and FATLBO algorithms.

Each benchmark function is tested 30 times with TLBO and FATLBO algorithms, and results are obtained in the form of mean value and standard deviation. The population of TLBO and FATLBO consists of 50 individuals. In this experiment, both algorithms are compared with DE variants. Except for the TLBO and FATLBO algorithms, the rest of the results are taken from those previous works. The comparative results of TLBO, FATLBO, and DE variants are shown in Table 10.

Table 10 Comparative results of FATLBO and other algorithms in Experiment 4

It can be seen from the results that the FATLBO algorithm has found 22 solutions that are better than others. Meanwhile, TLBO, DECC-G, DECC-C, and SaNSDE found only 16, 6, 7, and 0 solutions better than others, respectively. This shows that FATLBO is not only capable of handling low-dimension problems, but that it can also solve large-scale optimization problems. Meanwhile, FATLBO is able to outperform TLBO in F4, F5, F7, F12, and F13. For comparison purpose, the convergence graphics of TLBO and FATLBO algorithms for solving F13 are shown in Fig. 8. With the improved convergence characteristics provided by FATLS and a remedial operator, FATLBO is able to identify promising search regions faster than the regular TLBO.

Fig. 8
figure 8

Convergence graphics of TLBO and FATLBO algorithms for F13 with D = 500 and D = 1000

6 Conclusion

This paper develops a new advanced metaheuristic algorithm so-called fuzzy adaptive teaching–learning-based optimization (FATLBO). FATLBO presents three new strategies to enhance the performance of TLBO, namely a status monitor, fuzzy adaptive teaching–learning strategies (FATLS), and a remedial operator. Two new parameters, teaching rate (TR) and learning rate (LR), are proposed to present the probability rate for a student to enter teacher and learner phases. According to the mathematical benchmark function results, it is proven that the FATLBO has a better performance than the TLBO, the ABC, the DE, and the PSO, as well as their variants. When it was applied to benchmark problems, the FATLBO had the capacity to generate much better solutions regarding the global optimization than other metaheuristic algorithms. Based on the obtained results, we conclude that the FATLBO has the capacity to achieve more optimal results with less function evaluations than previously tested algorithms.

The integration of three new modifications into an existing TLBO algorithm adds huge advantages to the searching capability of TLBO. The status monitor measures the productivity of students during the teacher and learner phases. FATLS adapts the current teaching–learning strategy into either a more teacher-centered orientation or more student-centered orientation based on the productivity information collected from the status monitor. The remedial operator applies random perturbation using Gaussian distribution principle when no fitness improvement occurs for a long time or the entire population converges in one solution. The status monitor and FATLS increase the efficiency of knowledge transfer to each student and thus improve the optimization precision and convergence speed. Meanwhile, the remedial operator prevents the stagnation and avoids being trapped in local optima. We can conclude that the three new elements introduced here improve the quality of the results and increase the convergence speed of the basic TLBO algorithm. Thus, the novel FATLBO algorithm is proven to be a strong alternative for solving global optimization.

In addition, this paper defines a new understanding of parameter adaptation that may be described as one potential method for the parameter tuning process of FATLBO. More work is required to gather further evidence for possible shortcomings of the proposed algorithm. Meanwhile, optimizing the membership functions of each input variable can be a fruitful research agenda to further enhance the performance of the proposed algorithm. Finally, the application of the present fuzzy parameter adaptation to emerging metaheuristic algorithms like Jaya algorithm, SOS, and MVO could be a thought-provoking approach for future research.