1 Introduction

Scheduling is a decision-making process, perform vital role in services, manufacturing and production industry. It is traditionally defined as a process of allocating job sequence on different machines that reduces the makespan (completion-time) of a job sequence [1]. Scheduling problems are found in many real-world industries such as textile [2], discrete manufacturing industries [3], electronics [4], chemical [5], production of concrete [6], manufacturing of photographic film [7], iron and steel [8], and internet service architecture [9]. Typically, scheduling problems are classified on the basis of production environments such as single machines, flow-shop, parallel-machines, open-shop, cyclic flow shop and Flexible Job-Shop (FJS). PFSSP is an ultimate engaging research problem in the manufacturing industry with an objective to minimize the total completion time. It is one of the simplified versions of the typical Flow Shop Scheduling Problem (FSSP). The first PFSSP pioneering work was carried out on two machines, with a view to minimize the completion-time [10]. In terms of complexity estimates, PFSSP is found to be NP hard optimization problem [11].

The approaches to solve PFSSP can be broadly grouped into three categories: exact, heuristics and metaheuristic algorithms. The exact-algorithms: dynamic-programming [12], branch and bound techniques [13], and linear-programming [14], that can be successively adopted for solving small problems. However, these techniques do not produce promising outcomes for large instances in reasonable time. Therefore, heuristic techniques are generally used for large instances. The heuristic techniques are simple, fast and can be used to frame scheduling solutions [15]. A solution can be obtained in stipulated time using the heuristic algorithms, but the produces outcomes may not be optimal. Hence, metaheuristic algorithms are proposed over heuristic algorithms. An existing metaheuristic algorithms for any given problem provide balance between diversification and intensification [16]. There are many nature-inspired algorithms that are used to solve the PFSSP such as genetic algorithm [17], ant colony optimization [18], differential evolution [19], harmony search [20], firefly algorithm [21], bat algorithm [22], cuckoo search [23], TLBO [24], grey wolf algorithm [25], earthworm optimization algorithm [26], monarch butterfly optimization [27], artificial bee colony algorithm [28], jaya algorithm [29], moth search algorithm [30], return-cost-based binary FFA (Rc-BBFA) [31] etc. To overcome the problems faced by single metaheuristic algorithm, recently hybridization of algorithms i.e. combination of two or more local or global search algorithms are used. This also results in increased performance of the metaheuristic algorithms. As the hybridization algorithms can find high-quality feasible outcomes in reasonable time, they are recently used in research development of PFSSP. The advantages for selection of TLBO algorithm, it is easy to implement and easily applicable for practical application. Selection of parameter tune problem is not here. It is not required any mutation and crossover parameters as like genetic algorithm. The performance of TLBO is better than comparative other metaheuristic algorithms. The proper precaution have to take for TLBO algorithm as convergence rate is quickly that is main disadvantage of this algorithm.

In recent decades, researchers are focusing on the hybridization of the genetic algorithm in search of minimum makespan of PFSSP [32]. Hybrid Genetic Algorithm (HGA) [33] was proposed to solve sequence independence of PFSSP. The objective of Particle Swarm Optimization (PSO) for PFSSP, i.e. PSOVNS [34] was to minimize the makespan and the total-flow time. Also, the effective PSO with local search simulated annealing algorithm was used for PFSSP, which balance the exploitation and exploration [35]. Tabu-search-algorithm (TSA) hybridized with improved global search algorithm for solving PFSSP [36]. Hybrid Differential Evolution (HDE) algorithm combined with greedy based local search and Individual Improving Scheme (IIS) was developed to elevate the feasible outcomes quality [37]. The novel Hybrid Cuckoo Search (HCS) was proposed to minimize the makespan and total flow time for solving PFSSP [38]. Enhanced Migrating Bird Optimization (MBO) algorithm was also used to solve a scheduling problem [39]. For minimizing the total flowtime, distributed PFSSP was introduced [40]. While, Improved Migrating Birds Optimization (IMMBO) and a Hybrid-multi-objective Discrete Artificial-Bee-Colony (HDABC) were used to minimize the makespan [41]. Flow shop-scheduling problem to minimize makespan by modified fruit-fly optimization algorithm [42]

Teaching-Learning-Based Optimization (TLBO) was suggested by Rao et al. [24]. It was inspired from the teaching-learning process and it proved to be efficient. The main advantage of using TLBO is that, it is free from any algorithm-parameters. Due to its characteristics, the TLBO algorithm has been gaining recognition in the research industry. The electrochemical discharge machining technique and electrochemical-machining mechanism have been designed using TLBO algorithm [43]. The efficiency of TLBO algorithm was tested on flow-shop and Job-Shop-Scheduling Problem (JSSP) [44]. Shao et al. [45] applied hybrid TLBO algorithm combined with simulated annealing for solving PFSSP. The discrete TLBO algorithm (DTLBO) for solving flowshop rescheduling problem [46]. The probabilistic TLBO algorithm was used to solve no-wait flow-shop scheduling problem (NWFSSO) with high completion-time [47]. The flexible flow shop scheduling problems were solved using the improved TLBO and the JAYA algorithm [48]. The Whale-Optimization-Algorithm (WOA) with a local search approach was also used for solving PFSSP [49].

With this background, the primary contributions of this paper are listed below:

  1. 1.

    MTLBO with OBL method is applied for solving PFSSP. The OBL approach is used to enhance the quality of initial populations (individuals). Using Largest-Order-Value (LOV) rule based random key, MTLBO algorithm is developed to solve PFSSP effectively by changing individuals into discrete job sequences. Nawaz–Encore–Ham (NEH) heuristic method is combined with discrete job sequence to initialize a high quality and a diverse population. Moreover, to enhance the local exploitation ability, the effective swap, insert and inverse structures are included into the MTLBO algorithm.

  2. 2.

    In order to evaluate and analyze the effectiveness of the proposed algorithm, computational experiments are conducted using ten benchmark test functions and Wilcoxon signed-rank test.

  3. 3.

    Comprehensive extensive experiments are conducted using different well-known datasets such as Carlier, Reeves, Heller, Taillards and VRF benchmark test functions.

  4. 4.

    To validate its performance, the MTLBO algorithm is compared with various well-known metaheuristic algorithms.

The structure of the paper is as follows: Sect. 2 presents the description of the PFSSP. Section 3 provides the related work of basic TLBO algorithm. Section 4 describes the proposed algorithm MTLBO. Section 5 presents computational results and discussion and at last Sect. 6 discusses the conclusions. The symbols used to describe the PFSSP and along with their explanations are shown in Table 1.

Table 1 Symbols used

2 The description of the PFSSP

In the PFSSP, the set of n jobs \((j=j_1,j_2,\ldots ,j_n)\) are consecutively handled on the set of m machines \((M=M_1,M_2,\ldots ,M_m)\). As each machine can process only one job at a time, the handling time for each job is determined as \(P_{i,j}=(i=1,2,\ldots .,n,\) \(j=1,2,\ldots ,m)\). Each job is given a task or operation on each machine and is represented as \(T_j=T_{j1},T_{j2},\ldots ,T_{jm}\). The sequence followed by each job over each machine is the same. A schedule \(\varPi \) is denoted as a permutation \((\varPi =\pi _1,\pi _2,\ldots ,\pi _n)\), where \(\varPi \) represents set of all permutations for n jobs. Let us assume \(Z(\pi _i,m)\) to be the makespan of job (\(\pi _i\)) over the machine m. \(S_{1,\pi _i-1,\pi _i,}\) indicates initial time of job-permutation. Let us assume that \(P_{\pi _i,j}\) represents the handling time of job (\(\pi _i\)) over the machine j. Hence, the PFSSP can be mathematically represented as follows:

$$\begin{aligned}&Z\left( {{\pi }_{1}},1 \right) ={{p}_{{{\pi }_{1}},1}} \end{aligned}$$
(1)
$$\begin{aligned}&Z{(\pi _i,1)}=Z(\pi _{i-1},1)+p_{\pi _i,1}+S_{1,\pi _i-1,\pi _i,}\nonumber \\&\quad i=2,3,\ldots .,n \end{aligned}$$
(2)
$$\begin{aligned}&Z{(\pi _i,j)}=Z(\pi _1,j-1) +p_{\pi _1,j},~~~j=2,3,\ldots .,m \end{aligned}$$
(3)
$$\begin{aligned}&Z{(\pi _i,j)}=max (Z(\pi _{i-1},j)+S_{j,\pi _i-1,\pi _i,}\nonumber \\&\quad (Z(\pi _i,j-1)) +p_{\pi _i,1}~~ ~ \nonumber \\&\qquad i=2,3,\ldots .,n; ~~j=2,3,\ldots .,m \end{aligned}$$
(4)

where n indicates the total number of jobs, m denotes the machines, P represents PFSSP and \(Z_{max}\) is the goal to minimize makespan of the last job on the last machine.

In PFSSP, main goal is minimization of the maximum makespan, which can be determined as

$$\begin{aligned} {{Z}_{\max }}\left( \pi \right) =Z\left( {{\pi }_{n}},m \right) \end{aligned}$$
(5)

The main goal of PFSSP is to determine the optimal permutation from the set of all permutations:

$$\begin{aligned} Z\left( {{\pi }^{opt}} \right) \le Z\left( {{\pi }_{n}},m \right) \quad \text {for}\ \text {all}\ \varPi ~ \end{aligned}$$
(6)

where \(\pi ^{opt}\) denotes the optimal permutation.

3 Basic TLBO algorithm

Teaching-learning process is widely known since the ancient times, where one individual (learner) tries to learn from other individual. It has been applied in optimization of two thermoelectric coolers [50], cluster data [51] and image processing [52].

TLBO [24]is a population-based metaheuristic algorithm, that has two main fundamental phases: a teacher-phase and the learner-phase. In teacher phase of the algorithm, teacher teaches or shares their knowledge to learners (students). In the learning phase, the group of students or learners enhance their knowledge through teacher or interaction among themselves. Figure 1 illustrates the schematic flow of basic TLBO algorithm.

Fig. 1
figure 1

The schematic flow diagram of basic TLBO

3.1 Teacher-phase (TP)

It invokes the exploration phase of TLBO where students learn from the teacher. In this stage, teacher attempts to enhance the mean result of the class. For the objective function f(Z) with N-dimensional variable, \(Z=(z_{i,1},z_{i,2},\ldots ,z_{i,N})\) represents the position of ith learner (or student) solution for N-dimensional problem. Thus, the class mean position with P learners (where P is the size of population) is represented as

$$\begin{aligned} Z_{mean}=\frac{\sum _{i=1}^{P}z_{i,1},\sum _{i=1}^{P}z_{i,2},\ldots ,\sum _{i=1}^{P}z_{i,N}}{P} \end{aligned}$$
(7)

The students with the optimal value (i.e. best value) in current generation are represented as \(Z_{teacher}\). The position of each student is shown in Eq. (8)

$$\begin{aligned} {{Z}_{i,new}}={{Z}_{i,old}}+rand({{Z}_{teacher}}-{{T}_{F}}.{{Z}_{Mean}}) \end{aligned}$$
(8)

where \({Z}_{i,new}\) and \({Z}_{i,old}\) are the ith students (learners) new and old positions respectively. Rand is dispersed random number in the range of [0, 1]. \({T}_{F}\) is teacher-factor which decided average value to be altered. \({T}_{F}\) can be any 1 or 2, it can be generated randomly using Eq. (9) as

$$\begin{aligned} {{T}_{F}}=round[1+rand(0,1)\{2-1\}] \end{aligned}$$
(9)

where \({T}_{F}\) is achieved randomly when the TLBO in range of [1–2], where “1” represent to no raise in proficiency-level and “2” represents to entire transmission of knowledge.

3.2 Learner-phase (LP)

The learner-phase is an exploitation phase of TLBO algorithm, where a learner randomly collaborates with other learners to improve his/her proficiency. Student (Learner) \(X_i\), learns new things if another student \(X_{j} (j \ne i)\) has more proficiency than him/her. Learning process can be depicted mathematically as

$$\begin{aligned} {{Z}_{i,new}}={\left\{ \begin{array}{ll} {{Z}_{i,old}}+rand*\left( {{Z}_{i}}-{{Z}_{j}} \right) , &{} \\ \text {if}\quad f({{Z}_{i}})\le f({{Z}_{j}}).\\ {{Z}_{i,new}}={{Z}_{i,old}}+rand*\left( {{Z}_{i}}-{{Z}_{j}} \right) , &{} \\ \text {if}\quad f({{Z}_{i}})>f({{Z}_{j}}). \end{array}\right. } \end{aligned}$$
(10)

where rand is a random number in the range of [0, 1], \(Z_{i,new}\) it gives the preferable feasible value.

3.3 Duplication elimination

A duplicate solution should not be kept in the population as it may cause premature convergence of the algorithm [53, 54]. The duplication elimination strategy is given in Fig. 2.

Fig. 2
figure 2

Elimination of duplicate solution in TLBO algorithm

4 Proposed MTLBO approach

In this section, the proposed MTLBO is explained with the local search strategy for makespan minimising the makespan in PFSSP. The proposed algorithm flowchart is given in Fig. 3. The symbol and meaning of the proposed algorithm is shown in Table 2.

Table 2 Symbols and meaning used in proposed algorithm
Fig. 3
figure 3

Proposed MTLBO algorithm

4.1 Representation of solution for MTLBO algorithm

The basic TLBO algorithm was initially proposed for solving different continuous-optimization examples. However, the PFSSP is a discrete optimization technique. The basic TLBO algorithm, therefore cannot be used to solve PFSSP directly. So, the basic TLBO has to be modified using the concept of differential evolution with random scale factor (DERSF) [55]. This term is scaled by scale factor (R) in a random way in the range (0.5, 1) and it is given as

$$\begin{aligned} R=0.5(1+rand(0,1)) \end{aligned}$$
(11)

To apply MTLBO to PFSSP, one of the key concerns is to create mapping rule between job sequence and the vector of individuals. To convert individual \(Z_i=[z_{i,1}, z_{i,2},\ldots , z_{i,n}]\) into job permutation \(\varPi _i=[\pi _{i,1},\pi _{i,2},\ldots ,\pi _{i,n}]\), the following random keys (called as robust representation [56]) are used namely the Smallest Positive Value (SPV) [34], Largest Ranked Value (LRV) [57] and Largest Order Value (LOV) [58]. The SPV, LRV and LOV rules are used in existing research [38]. The job sequence relationship scheme plays a vital role in the developmental mechanism of the proposed MTLBO. If exchange of job sequence is not proper in PFSSP, then it will increase the computational time of the proposed algorithm.

In this study, LOV rule is utilized. LOV is a very simple method, which is focused on random-key presentation. In LOV rule, by ranking the individual of \(Z_i=[z_{i,1},z_{i,2},\ldots .,z_{i,n}]\) with decreasing order, temp job permutation sequence is generated as \(\sigma _i=[\sigma _{i,1},\sigma _{i,2},\ldots ,\sigma _{i,n}] \). Then a job sequence permutation is obtained as \(\varPi _i=[\pi _{i,1},\pi _{i,2},\ldots ,\pi _{i,n}]\) using Eq. (12).

$$\begin{aligned} {{\pi }_{i{{\sigma }_{i,j}}}}=j \end{aligned}$$
(12)

where \(j\in [1,\ldots .,n]\). In Table 3, LOV-rule is described with an easy example (n = 6) where the solution representation of individual vector (\(Z_i\)) is obtained with job dimension, position (location) and job permutation.

For \(Z_i\)= [0.6, 0.12, 1.8, − 0.48, 1.7, − 1.06], it can be seen that \(z_{i,3}\) has the largest value, and so it is selected as the first order of a job permutation. After that \(z_{i,5}\), \(z_{i,1}\), \(z_{i,4}\) and \(z_{i,6}\) are preferred as job-permutation. Hence, the job sequence \(\varPi _i\) = [3, 5, 1, 2, 4, 6] is obtained. From this formulation, the conversion using LOV rule is clear, which makes MTLBO suitable for solving PFSSP.

Table 3 Individual solutions represented in LOV-rule

4.2 Population initialization

This phase plays a vital role and it is applied uniformly and randomly. A population vector \(Z_i=[z_{i,1},z_{i,2},\ldots ,z_{i,n}]\) is randomly generated. Nawaz–Enscore–Ham (NEH) method is used for producing good initial population. NEH is the heuristic method to obtain optimal solution to the PFSSP. The NEH steps are described below:

  1. Step 1:

    All jobs are arranged in non-increasing procedure based on their overall handling time on all machines. Then job permutation \(\varPi (i)=[\pi (1),\pi (2),\ldots .,\pi (3)] \) is achieved.

  2. Step 2:

    Select any two jobs permutations, for examples \(\pi (1)\) and \(\pi (2)\). Then, all forms of permutation records of these two jobs are calculated and then an optimal order is taken.

  3. Step 3:

    Select any job permutation \(\pi (j)\),(j=3,4,....,n) and determine best permutation of jobs by assigning the available position to which they had scheduled. The optimal arrangement is selected from the next iteration.

The outcome created by NEH algorithm is a discrete job-permutation. To implement the MTLBO algorithm, the job permutation sequences are converted into individual by using Eq. (13), the conversion process is performed as:

$$\begin{aligned} \begin{aligned} Z_{NEH,i}&=Z_{min,i}+\frac{\left( Z_{max,i} -Z_{min,i} \right) }{n}\\&\quad \times \left( I_{NEH,i}-1 \right) ~~~~~~~ i=1,2,\ldots ,n \end{aligned} \end{aligned}$$
(13)

where \(I_{NEH,i}\) is job-index in ith dimension of job-permutation (\(\varPi \)). \(Z_{max,i}\) and \(Z_{min,i}\) are maximum and minimum bounds. \(Z_{NEH,i}\) is the individual value of given problem at the ith dimension.

Let us consider that all job-permutation (\(\varPi \)) achieved by NEH heuristic approach are given as \(\varPi \) = [2, 6, 3, 5, 1, 4]. Hence, the index of first carry on job sequence is given as value 5, index of second carry on job sequence is given as value 1. The index sequence (\(I_{NEH,i}\)) and individual value (\(Z_{NEH,i}\)) are shown in Table 4.

Table 4 Transformation of the job-permutation (\(\varPi \)) to the individual vector

4.3 Opposition based learning

Many optimization techniques are developed using OBL. OBL is used to enhance the performance and search ability to find a solution of population based algorithms [59]. Rahnamayan et al. [60] first proposed OBL strategy in an optimization method. The main purpose of OBL is to choose a better current candidate outcome by simultaneous evaluation of the current outcomes and its opposite outcomes. Mainly, OBL is applied to any population based algorithm during two phases: initial population and evolutionary phase. We can improve the initial population of proposed MTLBO algorithm using OBL. The concept of OBL depends on opposite number and opposite point.

Opposite number : Let a \(z \in [x, y]\) be any real number \(\mathbb {R}\). its opposite number \(z^{opp}\) is given by

$$\begin{aligned} z^{opp}=x+y-z \end{aligned}$$
(14)

Opposite point: Let \(Y(z_1,z_2,\ldots ,z_n)\) be point in N- dimensional space where \(z_i=[x,y];\). The opposite point of opposite population (\(z_{1}^{opp},z_{2}^{opp},\ldots ,z_{n}^{opp}\)) is given as:

$$\begin{aligned} z_{1}^{opp}=x_i+y_i-z_i \end{aligned}$$
(15)

4.3.1 GOBL optimization in MTLBO algorithm

The convergence rate of MTLBO algorithm can be improved with the help of GOBL. Wang et al. [61] proposed Generalized OBL model (GOBL). The proposed MTLBO approach uses this method to improve the performance of algorithm. When the premature convergence and stagnation state are identified, re-initialization method (i.e. OBL concept) is stimulated to avoid premature convergence in MTLBO algorithm.

4.4 Local search approach

The local search approach is applied to improve the quality of best scheduling (let \( B^*\)) such as

  1. (a)

    In the job permutation, each job in the \(B^*\) may have variable dimension d, where move the job sequence to all other (d − 1) location.

  2. (b)

    For each movement, calculate the new schedule of the job permutation, i.e. consider \(B^*\)= new schedule.

  3. (c)

    Least search probability (LSP) is used for local search approach.

By using the Kaveh et al. [62], the probability value P is changed to each generation(iteration) such as

$$\begin{aligned} P=0.3\left( \frac{iteration}{\max \_iteration} \right) \end{aligned}$$
(16)

where iteration is current iteration and \(max\_iteration\) is total number of iterations required to perform the experiments. Thus, changing the value of P improves the solution of the outcomes.

4.5 Swap, insert and inverse operators

To enhance the performance and quality of solution for MTLBO, some operators are chosen such as swap, insert and inverse. These operators would enhance local search ability [63] and improve the outcome solution. Let us randomly take any two different positions m and n in a job permutation (\(\pi \) ), if \(m <u\), m will be inserted at the back of u. Otherwise, m is to be inserted in front of u. The details of these structures are as follows:

4.5.1 Swap operator

The Fig. 4 shows the swap operation. Randomly select any two separate locations from a job-permutation (\(\pi \)) such as m and u and swap them.

Fig. 4
figure 4

Local-search using swap operator

4.5.2 Insert operation

In Fig. 5, randomly select any two separate locations from a job permutation. Here job permutation (\(\pi \)) is removed from position m and inserted into other position u.

Fig. 5
figure 5

Local-search using insert operator

4.5.3 Inverse operation

In Fig. 6, randomly change the subsequence between two separate locations of a job permutation (\(\pi \)). By performing the inverse operation, jobs in solution \(\pi \) and \(\pi _{new}\) are interchanged.

Fig. 6
figure 6

Local-search using inverse operator

4.6 Re-initialization structure

The experiment is conducted on the proposed algorithm, local search is used for fast convergence of the outcomes. Thus, at some extent the algorithm outcomes are trapped in local optimum i.e. premature convergence may occur. To overcome this situation, re-initialization mechanism is developed. If the feasible optimal solution is not found after 40 successive generation, then reinitialize the learners (populations). The re-initialization scenario is developed and it is given as: half of the populations are exchanged with the best populations achieved in the earlier process and another half of the populations are generated randomly.

4.7 Computational-complexity of MTLBO in term of big O notation

Let there be n jobs and m machines in PFSSP, the size of population is p. Generally in EA the computation complexity determine as of O\((d*p + fit*p)\), d is dimension of problem, fit is fitness or cost function (here minimization of maximum makespan is cost function). In the Initial stage of MTLBO computational complexity of NEH O() and remaining population generated by randomly using OBL technique is O(0.5 * \(n^2\) mp) . The computational complexity at initial phase is O(\(n^2mp\)) + O(\(0.5*n^2 mp\)). In teaching phase, due swap and local search computational complexity is O(\(n^2mp\)) . In learning phase, computational complexity is O(\(n^2m/2\)), due the learner gets information from teacher or exchange information themselves.

Total computation complexity of MTLBO for computing PFSSP :

O(nmp) = O(\(n^2mp\)) + O(\(0.5*n^2 mp\)) + O(\(n^2mp\)) + O(\(0.5*n^2 mp\))+ O(\(n^2m/2\)) \(\approx \)O(\(n^2mp\)) = O(50 \(\times \) 1000 \(n^2\)) = O(50000 \(n^2\))

Table 5 Parameter used for experiments of MTLBO algorithm

5 Computational results and comparisons

This section represents the computational outcomes generated from the MTLBO algorithm. The MTLBO is tested over ten standard benchmark test functions which are taken from Liang [64]. These test functions are either unimodal or multimodal. The tuning for various performance measures are calculated from various evolutionary algorithms by the best, worst, mean (average) and SD (standard-deviation) outcomes [65]. The unimodal benchmark functions validate accuracy and the convergence speed of proposed algorithm whereas multimodal benchmark test functions validate using the global search ability of proposed algorithm. Then, the computational results of MTLBO algorithm are analyzed using Wilcoxon signed rank test method. There are various kinds of stopping criteria in the literature of NIAs, like maximum-number of iterations (G) , number of function evaluation (NFEs), limit of time and tolerance. The experiment uses G (i.e. number of generations) as stopping criteria for proposed MTLBO algorithm.

Table 6 The comparison results of GA, PSO, DE, HS, TLBO, MTLBO (D = 50)
Table 7 Wilcoxon signed rank test for p-value comparing MTLBO with various EAs over ten benchmark functions (\(p\le 0.05\))
Table 8 Comparison of MTLBO algorithm on Carlier benchamark datasets

5.1 Experimental setup

In the following computational experiments, the performance and efficiency of MTLBO algorithm is applied on extensively used benchmark functions. There are mainly five well known benchmark functions such as

  1. (a)

    Carlier datasets (contains eight instances) consist of eight combinations, ranging from 11 jobs with 5 machines to 8 jobs with 8 machines [66].

  2. (b)

    Reeves and Yamada benchmark datasets (contains twenty-one instances) consist of seven combinations, ranging from 20-jobs with 5-machines to 75-jobs with 20-machines [67].

  3. (c)

    Heller datasets (contains two instances) consist of two combinations, ranging from 20 jobs with 10 machines to 10 jobs with 10 machines [68].

  4. (d)

    Taillard datasets (contains 120 instances) consist of twelve combinations, ranging from 20 jobs with 5 machines to 500 jobs with 20 machines [69].

  5. (e)

    VRF benchmark functions [70] (each contain 240 small and 240 large instances, which have upto 800 jobs on 60 machines) data is available at http://soa.iti.es.

The performance measure indicators for benchmark functions are ARPD (Average Relative Percentage Deviation ), BRPD (Best Relative Percentage Deviation ) and WRPD (Worst Relative Percentage Deviation) [71]. They are calculated in Eqs. (17), (18), and (19) respectively.

$$\begin{aligned} ARPD= & {} \frac{Z_{best}-Z_{opt}}{Z_{opt}}. 100 \end{aligned}$$
(17)
$$\begin{aligned} BRPD= & {} \sum \limits _{i=1}^{n}{\left( \frac{{{Z}_{i}}-{{Z}_{opt}}}{{{Z}_{opt}}} \right) \frac{1}{n}}\times 100 \end{aligned}$$
(18)
$$\begin{aligned} WRPD= & {} \frac{\left( {{Z}_{worst}}-{{Z}_{opt}} \right) \times 100}{{{Z}_{opt}}} \end{aligned}$$
(19)

where \(Z_{opt}\) is best makespan value, \(Z_{best}\), \(Z_{i}\) and \(Z_{worst}\) are makespan solutions obtained in tested algorithms. Smaller value of ARPD indicates that the algorithm are better.

In Table 5, the given parameters are calibrate for MTLBO by conducting several experiments. All the experimentation and analysis of MTLBO algorithm have been implemented on an Intel 3.40 GHz core i5 processor where MTLBO algorithm was programmed with Matlab 8.4 (R2014b) under Win7(x64) with 8 GB RAM.

Table 9 Comparison of MTLBO algorithm on Reeves benchmark datasets

5.2 Testing proposed algorithm using benchmark functions

In this experiment, ten unconstrained benchmark functions are tested using MTLBO algorithm for population size (\(NP= 50\)) and dimension (\(D=50\)). The performance of the MTLBO is calculated by 25 independent runs over the given ten benchmark functions. In Table 6, the effective results of MTLBO are shows in terms of performance measures such as best, worst, mean and SD values on different population based algorithms.

Table 10 Comparison of MTLBO algorithm on Heller benchmark datasets
Table 11 Results achieved by MTLBO and TLBO for Taillard benchmark datasets

5.2.1 Diversification and intensification of proposed algorithm

The unimodal optimization function (for e.g. Fun2 : Schwefels) is used to find the balance between diversification (exploration) and intensification (exploitation). In the Fig. 7 indicate the convergence of diversity of population with iterations for unimodal optimization function (i.e schwefels function). This is applied for five metaheuristic algorithms (i.e. GA, PSO, DE, TLBO and MTLBO). In the Fig. 7, the graph indicates that proposed algorithm MTLBO decrease gradually during the exploration, hence global optimal value obtained. The solution obtained during exploitation of algorithm is precise and stable.

Fig. 7
figure 7

Convergence graph of population diversity of unimodal function (fun2: Schwefels) for metaheuristic algorithms (D = 50)

5.2.2 Statistical analysis of proposed algorithm

To investigate the MTLBO algorithm, statistical procedure is applied to perform analysis such as parametric and non-parametric tests. The parametric test is based on assumptions made during the analysis of an experiment in computational intelligence. On the other hand, the non-parameter test is applied when the previous conclusions are not satisfied. Usually, this test deals with nominal or ordinal data, but it can also be applied to continuous data by performing the ranking based transformation.

Table 12 Comparison of MTLBO with other population based algorithms on basis of ARPD using Taillard Benchmark datasets

The non-parametric test is applied to investigate the proposed algorithm. There are two types of non-parametric tests such as pair-wise comparison and multiple comparisons. To perform analysis of the proposed algorithm, pair-wise comparison is applied to MTLBO. In pair-wise procedure, comparison is made between two methods (algorithms), by obtaining p-value from one another. Wilcoxon signed rank test is a powerful non-parametric test for the pair-wise procedure as given in [72].

In Table 7, results of Wilcoxon signed rank test for MTLBO algorithm, with different metaheuristic algorithms over ten benchmark functions are shown. The test is conducted using 25 independent runs of the proposed algorithm, which differs significantly from other algorithms. Table 7 records the p-value for ten unconstrained benchmark functions with \(R+\) and \(R-\) values. The \(R+\) value indicate a sum of rank in which the first algorithm outperform compare to the second algorithm. \(R-\) value indicate a sum of ranks second algorithm outperforms compared to the first algorithm. The non significant p-value is represented in bold in Table 7.

From the Table 7, it is clear that the performance of the MTLBO algorithm is significantly better than other metaheuristic algorithms. Functions Fun1, Fun2, Fun3, Fun5, Fun7 and Fun10 have significant p-value (\(p\le 0.05\)), in MTLBO versus DE, whereas functions Fun4, Fun6 and Fun9 does not have significant value.

5.3 Comparison of MTLBO algorithm on Carlier benchmark datasets

The effectiveness of MTLBO is compared with various metaheuristic algorithms on Carlier benchmark datasets. Table 8, shows Carlier datasets range from Car01 to Car08 with size range from \(11 \times 5\) to \(8 \times 8\). The performance of MTLBO is calculated using BRPD, ARPD and WRPD along with best makespan value (\(Z_{opt}\)). The MTLBO algorithm is compared with L-HDE [37], PSOVNS [34], DBA [73], HBSA [74], and PSOMA [35].

From the Table 8, it is observed that performance of proposed algorithm is better than given population based algorithm. The effective values of MTLBO algorithm shows better performance by using parameters of BRPD, ARPD, WRPD and best makespan values.

5.4 MTLBO algorithm on Reeves and Heller benchmark datasets

The effectiveness of proposed algorithm is compared with Reeves and Heller datasets. In the Table 9 indicates Reeves datasets range from REC001 to REC041 with size range from 20\(\times \) 5 to 75\(\times \) 20. The performance of MTLBO is calculated using BRPDARPD and WRPD along with best makespan value (\(Z_{opt}\)). The MTLBO algorithm is compared with L-HDE, PSOVNS, DBA, HBSA, PSOMA, and HCS [34]. Note that, the best values are shown in bold.

From Table 9, it is clear that the efficiency of MTLBO is superior than given population based algorithm. The effective values of MTLBO algorithm shows better performance by using parameters of BRPDARPDWRPD and best makespan values.

The comparison of the mean value of BRPDWRPD and ARPD for the dataset Reeves is shown in (Figs. 8, 9, 10). MTLBO is compared with various population-based algorithms, it can be seen that MTLBO algorithm achieves best outcomes. In mean value of BRPD, the largest value PSOVNS is 2.1665. In average value of ARPD, the largest value on DBA is 2.5875. In mean value WRPD, the largest value on DBA is 3.346. MTLBO shows best and effective outcomes as compared to other population based algorithms. Hence, MTLBO is most reliable and effective algorithm for solving PFSSP.

Fig. 8
figure 8

Comparison of mean value of BRPD on Reeves benchmark datasets

Fig. 9
figure 9

Comparison of mean value of ARPD on Reeves benchmark datasets

Fig. 10
figure 10

Comparison of mean value of WRPD on Reeves benchmark datasets

Table 13 Comparison of MTLBO algorithm with other evolutionary algorithms over Taillard benchmark problems
Table 14 Comparision of MTLBO on total flowshop time criterion over Taillard benchmark problems
Fig. 11
figure 11

Convergence plot of makespan over MTBLO and metaheuristic algorithms on Taillard functions

In Table 10, Heller datasets have two instances such as Hel1 and Hel2 with the size range \(20 \times 10\) and \(10 \times 10\). Hamdi et al. [75] determined the upper and lower bound of these datasets. The MTLBO is also compared with NEH, DSOMA [76] and SG [11]. It can be seen from the outcomes that MTLBO achieves better value for upper bound for Hel1 and Hel2 problems.

5.5 MTLBO algorithm on Taillard benchmark dataset

The comparison of the MTLBO with basic TLBO and other metaheuristic algorithms are carried out using Taillards benchmark functions [69]. The problem size of Taillard benchmark functions range from \(20 \times 5\) to \(500 \times 20\). The minimum and maximum bounds of Taillard’s test function are given in OR-Library ( i.e. Eric Taillard’s webpage).

5.5.1 Comparison of MTLBO with basic TLBO

To validate the performance of MTLBO, the experimental outcomes and comparison of the proposed algorithm with basic TLBO for Taillard benchmark test functions are shown in Table 11. The performance measure indicators such as BRPD, ARPD and WRPD are given in Table 11.

MTLBO algorithm outperforms the basic TLBO algorithm on all Taillard benchmark functions. The total average values for ARPD, BRPD and WRPD of MTLBO are 0.830, 3.717 and 6.118 which are better than basic TLBO algorithm tested values.

The convergence plots of basic TLBO and MTLBO algorithms are listed in Fig. 11. For plotting the graph, makespan versus Taillard benchmark function (such as TA010, TA030, TA050, TA070, TA090 and TA0110) are used. From these graphs, it can be indicated that the proposed algorithm converges faster than basic TLBO algorithm. As the iterations increase, MTLBO hardly finds accurate solutions, but by embedding the local search operator into MTLBO, it can achieve better performance.

Table 15 The p-value obtained by MTLBO versus other metaheuristic algorithms (\(p\le 0.05\))
Table 16 Comparison of MTLBO algorithm with other evolutionary algorithms over VRF hard small benchmark (t = 5 ms)
Table 17 Comparison of MTLBO algorithm with other evolutionary algorithms over VRF hard large benchmark (t = 5 ms)

5.5.2 Comparison of MTLBO with other EAs

In this subsection, the proposed MTLBO algorithm is compared with IG1 [77], IG2 [77], HDDE [78] and TLBO algorithms using Taillard benchmark test functions. The performance is measured using ARPD indicator. Table 12 shows comparison of proposed algorithm over population based algorithms with respect to ARPD. The MTLBO algorithm obtained lowest average value i.e. 0.83 for ARPD, which is better than IG1 (2.31), IG2 (2.27), HDDE (4.51) and TLBO (3.11).

Later on, the proposed MTLBO algorithm is compared with GA [79], HPSO [78], ATPPSO [80] and NCS [81] on the Taillard benchmark test functions. The computation results achieved by GA, HPSO, ATPPSO and NCS are indicated in Table 13. The performance measure indicators such as Mean and ARPD (Average relative error) are also presented in Table 13. The best outcomes of the benchmark functions are indicated in bold.

In Table 13, MTLBO obtains the best computational outcomes with respect to all the given metaheuristic algorithms. The MTLBO algorithm obtained lowest total mean ARPD (i.e. 0.8) which is superior value than HPSO (2.74), ATPPSO (1.94), GA (4.31), and NCS (1.24). The performance of proposed algorithm significantly superior outcomes over Taillard benchmark test function. The mean values of proposed algorithm are within the lower and upper bound given as OR-Library.

In Table 14, the proposed algorithm compared total flowshop time criterion over Taillard benchmark problems which taken from http://www.cs.colostate.edu/sched/generator. MTLBO compared with PSO and RSA [82] . The mimumium and average results obtained by MTLBO is significantly superior than PSO and RSA.

To analyze the performance of MTLBO algorithm over other evolutionary algorithms, Wilcoxon sign ranked test was conducted by using achieved ARPD value. Table 15 shows significant p-value of proposed algorithm over other evolutionary algorithms. From these experimental outcomes, it is observed that MTLBO algorithm is more effective and significant than GA, HPSO, ATPPSO and NCS. It is to be noted that the p-value in this experiment is shown in bold letters for below 0.05 value.

5.6 MTLBO algorithm on VRF benchmark datasets

New benchmark test functions such as VRF_hard_small benchmark and VRF_hard_large benchmark are proposed in [70]. These benchmark functions are selected to perform experiments on PFSSP. These functions consist of 240 small instances and 240 large instances, with upto 800 jobs on 60 machines. The small instances have following combinations of n = number of jobs, m = number of machines i.e. n = [10, 20, 30, 40, 50, 60] and m = [5, 10, 15, 20] and large instances consist of n = [100, 200, 300, 400, 500, 600, 700, 800] and m = [20, 40, 60].

The stopping criterion to perform VRF benchmark function is evaluated with maximum running CPU time as follows

$$\begin{aligned} Maximum ~ running~ CPU~time =(n.(m/2). t) \end{aligned}$$
(20)

where t indicates the amount of time (t = 5 ms). Each of the instance runs at least 25 times to obtain the best outcomes.

To measure the effectiveness of the metaheuristic method, ARPD is calculated for each instance as in Eq. (17).

The computational results of the MTLBO over different metaheursitic algorithms with VRF small instance are given in the Table 16. The best outcomes of the experimental are shown in bold letters. The performance of MTLBO is calculated using ARPD and SD values. In Table 16, the average value for MTLBO with VRF small instances for ARPD and SD are 0.10 and 0.03 respectively. From Table 16, it is observed that ARPD and SD value of MTLBO is superior than DPSO, GA-VNS, HDE and HPSO algorithms.

The computational results of MTLBO over different metaheursitic algorithms with VRF Large instances are given in the Table 17. The best outcomes of the experiments is shown in bold letters. The performance of MTLBO algorithm is calculated using ARPD and SD values. In Table 17, the average value of the MTLBO with VRF small instances for ARPD and SD are 2.61 and 0.14 respectively. From Table 17, it is observed that obtained results using ARPD and SD value for MTLBO algorithm are statistically better than DPSO, GA-VNS, HDE and HPSO algorithms.

6 Conclusion

This paper proposed a Modified Teaching Learning Based Optimization with Opposite Based Learning approach to find solution of Permutation Flow-Shop Scheduling Problem. OBL approach is used to enhance the quality of the initial population and convergence speed. MTLBO is developed to determine the PFSSP efficiently using the Largest Order Value rule-based random-key, to change an individual into discrete job schedules. Based on Nawaz–Enscore–Ham heuristic mechanism, the new initial populations are generated in MTLBO. To enhance the local exploitation ability, the effective swap, insert and inverse structures are incorporated into the MTLBO.

The statistical result and analysis of the MTLBO based on ten benchmark functions and Wilcoxon signed rank test shows the effectiveness of the algorithm. The computational results over five well-known benchmark functions such as Carlier, Reeves, Heller, Taillard and VRF benchmark instances set indicates that MTLBO is the most powerful and convenient method to solve PFSSP. The performance measures of the proposed algorithm are calculated using ARPD, BRPD, and WRPD over all datasets. The performance of MTLBO is tested against state-of-the-art algorithms. It is observed that MTLBO algorithm works effectively in most of the cases. Thus, by conducting Wilcoxon signed test, it is proved that MTLBO algorithm is much more effective than other metaheuristic algorithms. The future research direction focus on hybrid flow shop scheduling and multi-objective flow shop problems.