1 Introduction

Production and distribution are two fundamental and interrelated operations in supply chain management. To improve the operational performance of supply chains, it is critical to schedule these two operations in an integrated manner, which has received much attention from both industrial and academic communities [1, 29, 38]. In the integrated production and distribution systems, the manufacturer firstly processes the customer orders on the machines, and then timely distributes the finished products back to the customers [13, 14, 20]. To reduce the transportation cost from the manufacturer to its customers, the finished products are generally grouped into batches for delivery. A batch is a set of orders or finished products that are delivered in the same vehicle [47, 49]. However, to meet customer’s requirements under limited warehouse capacity, products need to arrive at the customers within tight delivery time windows, inevitably requiring frequent delivery of small batches and eventually leading to high transportation cost [22, 42]. Therefore, a trade-off between customer service and delivery performance has to be considered when integrating production and distribution decisions.

The hybrid flow shop (HFS) is a common manufacturing system to produce identical or similar products, especially in industries such as automobile, steel, electronics, semiconductors, textiles and food [39, 41]. In a typical HFS, machines are arranged into a number of stages in series, each of which has several functionally identical machines. All jobs released to an HFS have to pass through all stages in the same order [44]. To enhance competitiveness and responsiveness in rapidly changing markets, factories have shifted from a centralised to a more decentralized structure [5, 36]. Although an increasing research interest has recently focused on production scheduling in multi-factory environments, relatively few studies have been devoted to addressing the distributed HFS scheduling problems (DHFSSP). Given that the DHFSSP has been proven NP-hard in nature [54], integration of production and distribution scheduling in distributed HFS environments further aggravates its complexity.

As an increasing number of manufacturers adopt the make-to-order (MTO) policy to response to changing market demand, considerable amount of research works have been devoted to the integrated production and distribution scheduling problems (IPDSP) in a variety of manufacturing environments [11, 12, 33], mainly including single machine, parallel machine, and flow shop systems. The earliest research on IPDSP was conducted by Potts [37]. He considered a single machine scheduling problem with non-identical job release times and delivery times. Chang et al. [9] addressed a novel supply chain scheduling problem, in which the production and distribution of products were modelled as an identical parallel machine scheduling problem and a capacitated vehicle routing problem respectively. Considering the speed difference between machines, Guo et al. [21] focused on an unrelated parallel machine scheduling problem with batch delivery. In this problem, several transportation modes with different vehicle capacities and transportation times were considered. Hassanzadeh et al. [24] investigated a multi-objective supply chain scheduling problem, in which jobs are firstly processed in a permutation flow shop and then transported to the customers in batches with unlimited capacity. More recently, to deal with IPDSP in the assembly flow shop environments, Kazemi et al. [30] and Basir et al. [4] developed effective approaches to coordinate production and distribution operations. A search of available literature on IPDSP indicates that no research works have attempted to integrate production and distribution scheduling in a distributed HFS environment.

Exact algorithms, heuristics and meta-heuristics are three commonly used approaches to address the IPDSP in the literature. Exact algorithms, such as the dynamic programming [9, 23], branch-and-bound approach [32, 43], and branch-and-price approach [3], have been applied to generate the optimal solution for small-sized instances. However, owing to high complexity of IPDSP, exact algorithms are inapplicable to deal with medium-sized and large-sized instances in industrial practice. Since heuristics and meta-heuristics are capable of obtaining near-optimal solutions within reasonable computational time [2, 7, 8], recent research efforts have shifted from developing exact algorithms to efficient heuristics and meta-heuristics, mainly including genetic algorithm (GA) [4, 24, 28], artificial immune system (AIS) [26], particle swarm optimization (PSO) [40], ant colony optimization (ACO) [13, 14, 19], imperialist competitive algorithm (ICA) [30, 31], iterated greedy algorithm [52], and hybrid approaches [27, 46, 55].

This paper aims to address a novel IPDSP, in which products are first produced in a number of distributed HFSs and then delivered to a customer in batches. To satisfy the production and distribution practice, factory eligibility and multiple transportation modes are considered in the studied problem. In many real-world manufacturing systems, a job can only be processed on a subset of machines owing to some technical or physical reasons [17, 39]. This constraint, known as machine eligibility in the literature, has recently received much attention in the parallel machine and HFS scheduling problems [6, 56]. As the extension of machine eligibility, factory eligibility is considered in the studied IPDSP, which does not allow a job to be assigned to a factory if it cannot be processed on any parallel machines at some stage in the factory [53]. In addition, as many manufacturers have difficulty in providing sufficient transportation facilities to meet the customer’s demand, finished products are allowed to be delivered by both the manufacturer and the third-party logistics (3PL) provider. Therefore, two types of transportation modes, namely manufacturer transportation and 3PL transportation, are considered in the studied problem. The transportation cost by a vehicle of the manufacturer is fixed, whereas the transportation cost by a vehicle of the 3PL provider consists of fixed charge cost and a load-based transportation cost. To our best knowledge, it is the first attempt to integrate production and distribution scheduling in distributed HFSs with consideration of factory eligibility and multiple transportation modes.

Owing to the NP-hardness of the studied IPDSP, three fast heuristics (CR-based, SLACK-based, and EDD-based heuristics) and an adaptive human-learning-based genetic algorithm (AHLBGA) are proposed to obtain sub-optimal solutions in this paper. Each of the three heuristics integrates a simple scheduling rule with a job re-assignment method and a vehicle selection method to generate integrated production and distribution schedules. The uniqueness of the proposed AHLBGA is characterized by hybridizing GA with an adaptive human-learning operator. Apart from commonly used genetic operators of selection, crossover, and mutation, learning is another important evolutionary process to acquire new knowledge for enhancing competitive advantages of species. To improve the performance of offspring individuals after crossover and mutation, AHLBGA applies a novel learning operator to mimic human learning behaviours. This learning operator consists of three typical activities, namely social learning, family learning, and individual random learning. Furthermore, to provide a better trade-off between exploration and exploitation, the selection probability of different learning activities is dynamically determined by an adaptive learning scheme. A comprehensive search of literature on IPDSP indicates that no research works have attempted to incorporate GA and human learning activities to generate integrated schedules. Since distributed HFSs are typical production systems in many MTO industries, the proposed AHLBGA may help operations managers to better coordinate the scheduling of production and transportation.

The rest of paper is organized as follows. Considering factory eligibility and multiple transportation modes, Sect. 2 formulates a novel IPDSP in a distributed HFS environment. Section 3 presents three fast heuristics and AHLBGA to generate integrated production and distribution schedules. To evaluate the performance of proposed algorithms, experimental results on small-sized and large-sized problems are presented in Sect. 4. Finally, Sect. 5 summarises the research findings and identifies some future research directions.

2 Mathematical model

2.1 Problem description and assumptions

The IPDSP in distributed HFSs is a novel combinatorial optimization problem. As indicated in Fig. 1, this problem consists of two inter-dependent sub-problems, namely distributed HFS scheduling with factory eligibility in the production stage and batch delivery of finished products with two types of transportation modes in the distribution stage.

Fig. 1
figure 1

The distributed HFSs with batch delivery

In the production stage, a set J = {1, 2, …, n} of n jobs are released to f distributed identical factories for processing. Each of the factories is arranged as an HFS, which has a set S = {1, 2, …, t} of t stages and mk parallel machines at each stage k. The transportation time between stages in a factory is negligible and therefore not considered in this study. Owing to the technical or physical aspects, factory eligibility is considered in the production stage, in which some jobs can only be processed in a subset of distributed factories.

In the distribution stage, the finished products in each of the distributed factories are grouped into batches and delivered to a customer by the vehicles of the manufacturer and the 3PL provider. The manufacturer only provides a limited number of vehicles for transportation, whereas the 3PL provider offers a sufficient number of vehicles to meet the customer’s demand. The transportation cost by a vehicle of the manufacturer is fixed. The transportation cost by a vehicle of 3PL provider consists of a fixed charge cost and a load-based variable cost, which depends on the quantity of products in a batch.

To address the studied IPDSP, four operational decisions have to be made: (1) assigning jobs to suitable factories with consideration of factory eligibility; (2) scheduling the assigned jobs for each of the factories; (3) allocating the finished products into an appropriate number of batches; (4) assigning each of the product batches to a vehicle of the manufacturer or 3PL provider for delivery.

To simplify the integration of production and distribution, the following assumptions have been made: (1) all jobs are available for processing at time zero; (2) some jobs can only be handled in some specific factories; (3) factory switching is not allowed, that is, once a job is assigned to a certain factory, all its operations have to be performed in this factory; (4) each machine can process only one job at a time, and job preemption is not permitted; (5) the same job processed on any parallel machine at a stage takes equal processing time; (6) a batch is delivered immediately to the customer when all the jobs of the batch are finished; (7) the transportation time from different factories to the customer is fixed.

2.2 Problem formulation

Considering factory eligibility and multiple transportation modes, this section presents the mathematical model of the studied IPDSP. The parameters and decision variables used in the mathematical model are defined as follows:

2.2.1 Parameters

i :

index of factories, 1 ≤ i ≤ f.

j :

index of jobs, 1 ≤ j ≤ n.

k :

index of stages in a factory, 1 ≤ k ≤ t.

l :

index of parallel machines at a stage, 1 ≤ l ≤ mt.

b :

index of batches, 1 ≤ b ≤ p.

f :

the number of factories.

n :

the number of jobs released to distributed HFSs.

n i :

the number of jobs processed in the HFS of factory i.

t :

the number of stages in an FFS.

m t :

the number of parallel machines at stage t.

\(\pi_{i}^{j}\) :

jth job of the job sequence \(\pi_{i} = \left\{ {\pi_{i}^{1} ,\pi_{i}^{2} ,...,} \right.\left. {\pi_{i}^{{n_{i} }} } \right\}\) processed in factory i.

\(P(\pi_{i}^{j} ,k)\) :

processing time of job \(\pi_{i}^{j}\) at stage k in factory i.

\(TT_{i}^{{}}\) :

transportation time between factory i and the customer.

\(MTC_{i}^{{}}\) :

transportation cost of factory i by a vehicle of the manufacturer.

\(FTC_{i}^{{}}\) :

fixed transportation cost of factory i by a vehicle of the 3PL provider.

\(VTC_{i}^{{}}\) :

variable transportation cost of factory i by a vehicle of the 3PL provider.

\(DC_{b}^{{}}\) :

delivery cost of batch b.

\(d_{j}^{l}\) :

the earliness due date of job \(\pi_{i}^{j}\).

\(d_{j}^{u}\) :

the latest due date of job \(\pi_{i}^{j}\).

\(TC\) :

unit tardiness cost.

\(EC\) :

unit earliness cost.

2.2.2 Decision variables

\(\alpha_{{iklj_{1} j_{2} }}\) :

1 if job \(\pi_{i}^{{j_{2} }}\) is processed immediately after job \(\pi_{i}^{{j_{1} }}\) on machine l at stage k in factory i; 0 otherwise.

\(\eta_{ib}\) :

1 if the products of batch b are manufactured in factory i; and 0 otherwise.

\(\lambda_{ijb}\) :

1 if job \(\pi_{i}^{j}\) is assigned to batch b; and 0 otherwise.

\(\mu_{b}\) :

1 if batch b delivered by a vehicle of the manufacturer; and 0 otherwise.

\(p\) :

the number of batches delivered to the customer.

\(C(\pi_{i}^{j} ,k)\) :

completion time of job \(\pi_{i}^{j}\) at stage k in factory i.

\(DB_{b}^{{}}\) :

delivery start time of batch b.

\(DJ_{ij}^{{}}\) :

delivery start time of job \(\pi_{i}^{j}\) processed in factory i.

The paper aims to generate integrated production and distribution schedules in a distributed HFS environment. Based on the above notations, the studied IPDSP is formulated as the following mathematical model:

$$ Min\;\;\sum\limits_{i = 1}^{f} {\sum\limits_{j = 1}^{{n_{i} }} {TC \times max\left\{ {0,DJ_{ij} - d_{j}^{u} } \right\}} } + \sum\limits_{i = 1}^{f} {\sum\limits_{j = 1}^{{n_{i} }} {EC \times max\left\{ {0,d_{j}^{l} - DJ_{ij} } \right\}} } + \sum\limits_{b = 1}^{p} {DC_{b} } $$
(1)

Subject to:

$$ C(\pi_{i}^{j} ,1) \ge P(\pi_{i}^{j} ,1),\quad i = 1, \ldots ,f;j = 1, \ldots ,m_{1} $$
(2)
$$ C(\pi_{i}^{{j_{2} }} ,1) \ge \sum\limits_{l = 1}^{{m_{1} }} {\sum\limits_{{j_{1} = 1}}^{{n_{i} }} {\alpha_{{i1lj_{1} j_{2} }} } } C(\pi_{i}^{{j_{1} }} ,1) + P(\pi_{i}^{{j_{2} }} ,1),\quad i = 1, \ldots ,f;\,j_{2} = m_{1} + 1, \ldots ,n_{i} $$
(3)
$$ C(\pi_{i}^{j} ,k) \ge C(\pi_{i}^{j} ,k - 1) + P(\pi_{i}^{j} ,k),\quad i = 1, \ldots ,f;j = 1, \ldots ,m_{1} ;\,k = 2, \ldots ,t $$
(4)
$$ C(\pi_{i}^{{j_{2} }} ,k) \ge max\left\{ {\sum\limits_{l = 1}^{{m_{k} }} {\sum\limits_{{j_{1} = 1}}^{{n_{i} }} {\alpha_{{iklj_{1} j_{2} }} } } C(\pi_{i}^{{j_{1} }} ,k),C(\pi_{i}^{{j_{2} }} ,k - 1)} \right\} + P(\pi_{i}^{{j_{2} }} ,k),\quad i = 1, \ldots ,f;\,j_{2} = m_{1} + 1, \ldots ,n_{i} ;\,k = 2, \ldots ,t $$
(5)
$$ \sum\limits_{b = 1}^{p} {\lambda_{ijb} = 1} ,\quad i = 1, \ldots ,f;j = 1, \ldots ,n_{i} $$
(6)
$$ DC_{b} = \sum\limits_{i = 1}^{f} {\eta_{ib} max\left\{ {\mu_{b} MTC_{i} ,(1 - \mu_{b} )(FTC_{i} + \sum\limits_{j = 1}^{{n_{i} }} {\lambda_{ijb} VTC_{i} } )} \right\}} ,\quad b = 1, \ldots ,p $$
(7)
$$ DB_{b} { = }\sum\limits_{i = 1}^{f} {\eta_{ib} } max\left\{ {\lambda_{ijb} C(\pi_{i}^{j} ,k)} \right\},\quad j = 1, \ldots ,n_{i} ;\,k = 1, \ldots ,t;\,b = 1, \ldots ,p $$
(8)
$$ DJ_{ij} = \sum\limits_{b = 1}^{p} {\lambda_{ijb} DB_{b} } ,\quad i = 1, \ldots ,f;\,j = 1, \ldots ,n_{i} $$
(9)
$$ C(\pi_{i}^{j} ,k),DC_{b} ,DB_{b} ,DJ_{ij} > 0,\quad i = 1,2, \ldots ,f;\,j = 1, \ldots ,n;\,k = 1, \ldots ,t;\,b = 1, \ldots ,p $$
(10)

In the presented model, the objective function (1) is to minimize the sum of earliness, tardiness and delivery costs. For the first stage in a factory, constraints (2) and (3) give the completion time of the first job and the successive jobs on any parallel machine, respectively. Similarly, for all other stages in a factory, constraints (4) and (5) provide the completion time of the first job and the successive jobs on any parallel machine, respectively. Constraint (6) guarantees that each job assigned to a factory is allocated to only one batch for delivery. Considering multiple transportation modes, constraint (7) determines the delivery cost of a batch using the vehicles of the manufacturer or the 3PL provider. Constraint (8) ensures that a batch is delivered immediately after all its jobs have been completed. Constraint (9) determines the delivery start time of any job processed in a factory. Constraint (10) guarantees that all the time-related and cost-related decision variables are positive.

3 Algorithms for IPDSP in distributed HFSs

Considering the NP-hard nature of the studied IPDSP, it is infeasible to generate the optimal schedule in reasonable time, especially for large-sized instances. Therefore, this section presents three simple heuristics (CR-based, SLACK-based, and EDD-based heuristics) and a novel adaptive human-learning-based GA (AHLBGA) to obtain near-optimal solutions for integrated production and distribution scheduling in distributed HFSs.

3.1 Heuristics

Heuristics adopt simple strategies to construct solutions and therefore tend to be much faster but less effective. The most well-known scheduling heuristics are dispatching rules. Among a variety of dispatching rules, the critical ratio (CR), minimum slack (SLACK), and earliest due date (EDD) have been found more effective for earliness and tardiness related criteria [10, 35, 57]. To address the IPDSP with factory eligibility and 3PL distribution, three fast heuristics, namely CR-based, SLACK-based, and EDD-based heuristics, are developed to allocate jobs into an appropriate number of batches for production and distribution.

Considering factory eligibility and multiple transportation modes, each of the proposed heuristics integrates a simple scheduling rule with a job re-assignment method and a vehicle selection method. In the CR-based heuristic, the jobs released to the manufacturer are firstly sequenced in ascending order of CR, which is defined as the ratio of the remaining time until the due date to the total processing time in a factory. To reduce the search space of feasible solutions, the jobs are then evenly grouped into a number of batches (p) for production and distribution. As a batch contains at least one job, the maximum number of batches is limited to be the number of jobs n. Therefore, all possible numbers of batches, i.e.\(1 \le p \le n\), need to be examined in this heuristic. Following the obtained job sequence, each of the batches is assigned to a factory with the minimum completion time. For the jobs that cannot be processed in the assigned factory, a job re-assignment method is applied to re-allocate them to eligible factories. In addition, to reduce the transportation cost to the customer, a vehicle selection method is employed to allocate each of the finished batches to a vehicle of the manufacturer or 3PL provider for delivery. Owing to the limited number of manufacturer’s vehicles, this method applies the transportation cost difference under multiple transportation modes to determine the batches delivered by the manufacturer’s vehicles.

The proposed CR-based heuristic, the job re-assignment method, and the vehicle selection method are detailed as follows. In the CR-based heuristic, \(\left\lfloor x \right\rfloor\) indicates the greatest integer that is not more than \(x\).

figure a
figure b
figure c

To address the IPDSP with factory eligibility and 3PL distribution, the SLACK-based and EDD-based heuristics are also developed. These two heuristics employ the same procedure of the CR-based heuristic to generate the schedules, except that the jobs released to the distributed HFSs are processed in ascending order of \(\frac{{\left( {d_{j}^{l} { + }d_{j}^{u} } \right)}}{2} - \sum\nolimits_{k = 1}^{t} {P\left( {j,k} \right)}\) and the due time respectively.

3.2 Proposed AHLBGA

3.2.1 Framework of AHLBGA

To address the IPDSP with factory eligibility and 3PL distribution, AHLBGA is proposed to generate better solutions. As one of the most widely used population-based meta-heuristics, GA has been successfully applied to generate near-optimal solutions to a variety of complex optimization problem [25, 48, 51], including the HFS scheduling problem and IPDSP [47, 49]. It produces the offspring chromosomes by performing three genetic operators, namely selection, crossover, and mutation. Apart from the traditional genetic operators, learning is another crucial biological process that enables all the living organisms to adapt to the changing environment by applying their previous or newly acquired knowledge and skills [34, 47, 49]. However, the existing GAs in the literature seldom perform learning activities to improve the offspring individuals generated after crossover and mutation [16, 18].

Different from traditional GAs, the proposed AHLBGA integrates an adaptive learning operator with traditional genetic operators to obtain better integrated production and distribution schedules. In the AHLBGA, an initial population is firstly randomly generated, and then evolves through selection, crossover, mutation, and adaptive learning until the predefined computation time is reached. To imitate the human learning behaviours, adaptive learning involves three types of learning operations, namely social learning, family learning, and individual random learning. To enhance the learning performance of offspring individuals, the probability of selecting a learning operation is dynamically adjusted by an adaptive learning scheme (to be detailed in Sect. 3.2.5). Such scheme encourages more individual random learning along with population evolution to help escape from local optimum, and therefore leads to potentially better solutions. It is the first attempt in integrating GA with human learning process for the IPDSP. The framework of AHLBGA is shown in Fig. 2 and each part of AHLBGA is detailed in the following subsections.

Fig. 2
figure 2

Framework of AHLBGA

3.2.2 Encoding and decoding scheme

The encoding scheme plays an important role in designing an effective and efficient meta-heuristic. Different from traditional flow shop scheduling problems, the IPDSP needs to be solved by integrating the decisions of production and distribution. Based on the commonly used permutation encoding scheme for production scheduling, a matrix of two rows and n columns is employed to represent a solution (chromosome) of the IPDSP. In the matrix, the first row is a permutation of n jobs that indicates the processing order in a distributed HFS environment, whereas the second row shows the assigned batches of all jobs for production and distribution. Since a batch contains at least one job, each of the rows is a string of integer numbers from 1 to n.

Figure 3 presents a chromosome of an IPDSP with 10 jobs to be scheduled in the HFSs of two distributed factories. As described in the first row of the chromosome, jobs are suggested to be processed according to the order of {4, 1, 8, 6, 2, 9, 3, 10, 7, 5}. According to the batch numbers in the second row of the chromosome, these jobs are grouped into three batches, namely jobs {6, 10, 7, 5} in batch 1, jobs {4, 1, 3} in batch 2, and jobs {8, 2, 9} in batch 3.

Fig. 3
figure 3

A chromosome of proposed AHLBGA

Considering factory eligibility and multiple transportation modes, a new decoding procedure is employed to convert the matrix of a solution into a feasible integrated schedule. As shown in Fig. 4, this decoding procedure mainly consists of four steps: (1) apply the ECT rule to assign each of the batches to a factory with the minimum completion time; (2) identify the jobs that cannot be processed in the assigned factories, and re-assign them to other suitable batches using the job re-assignment method; (3) the batches assigned to the same factory are scheduled based on the FCFS rule, and the jobs in a batch are processed in parallel based on their order in the batch; (4) apply the vehicle selection method to determine the appropriate vehicles of the manufacturer or 3PL provider to deliver the finished batches in the production stage to the customer. The details of the job re-assignment method and the vehicle selection method have been presented in Sect. 3.1.

Fig. 4
figure 4

An example of the decoding procedure

Figure 4 describes the decoding process of the chromosomes shown in Fig. 3. According to the ECT rule, batches 1 and 3 are allocated to the FFS of factory 1, and batch 2 is assigned to the FFS of factory 2. Assuming that job 8 in batch 3 can only be processed in the FFS of factory 2, the constraint of factory eligibility is violated and job 8 is therefore re-assigned to batch 2, which is processed in the FFS of factory 2. After the job re-assignment, job processing in factories 1 and 2 follows the job sequence {6, 10, 7, 5, 2, 9} and {4, 1, 8, 3}, respectively. To reduce the transportation cost to the customer, finished batches in the production stage are assigned to suitable vehicles for delivery, i.e. batches 1 and 2 delivered by the manufacturer and batch 2 delivered by the 3PL provider.

3.2.3 Initialisation, evaluation, and selection

The chromosomes in the initial population are randomly generated to cover the entire search space.In each of the generations, the chromosomes are decoded by the proposed decoding procedure and evaluated using the following fitness function:

$$ f(x_{i} ) = {1 \mathord{\left/ {\vphantom {1 {y(x_{i} )}}} \right. \kern-\nulldelimiterspace} {y(x_{i} )}},\quad i = 1,2, \ldots ,PS $$
(11)

where \(x_{i}\) represents the ith chromosome of a population, \(y(x_{i} )\) denotes the objective value of \(x_{i}\), and \(PS\) is the number of chromosomes in a population.

Based on the above fitness function, the well-known roulette wheel selection method is employed to choose better chromosomes for reproduction. In this method, the probability of a chromosome to participate in the reproduction process is proportional to its fitness. To avoid the loss of the best chromosome, the elitist strategy is applied to preserve the best chromosome to the next generation. In this strategy, the chromosome with the lowest fitness is replaced by the best chromosome found so far.

3.2.4 Crossover and mutation

Once the chromosomes have been selected for reproduction, crossover and mutation are performed to generate the offspring by mimicking the evolution process in the nature. Crossover exchanges and recombines genetic information of parent chromosomes to produce better offspring [15], and mutation slightly perturbs the genes of selected offspring to maintain the diversity of populations [50].

As the two-point crossover has been proven to provide better performance for scheduling problems, it is adopted in the AHLBGA to produce new offspring. As described in Fig. 5, two selected parents are firstly divided into three parts by two random cut-off points. The genes in the second part (middle part) of the two parents are subsequently interchanged. Finally, the jobs and their assigned batches in the first and third parts are re-arranged according to their order in the other parent.

Fig. 5
figure 5

An example of two-point crossover

Considering the matrix structure of the chromosomes, a new mutation method is employed to enhance the local search and avoid trapping into local optimum. As shown in Fig. 6, this proposed mutation involves two separate operations, namely the swap mutation and the single-point mutation. The swap mutation chooses two random columns and swaps both the jobs and their assigned batches in these two columns, whereas the single-point mutation only select one random job and randomly re-assigns it to another batch.

Fig. 6
figure 6

An example of mutation

3.2.5 Adaptive learning

To improve the offspring individuals after crossover and mutation, AHLBGA applies a novel adaptive learning operator to construct better solutions. By mimicking human learning behaviours, this learning operator consists of the following three types of learning operations.

  • Social learning operation

The humans in the social environment gradually accumulate their knowledge and experience by direct or indirect interactions ([45]; Wang et al., 2017). Social learning from a group of knowledgeable and experienced individuals is an efficient way to develop the learner’s ability. To perform the social learning operation, an elite social set \(\Omega_{S}\) is firstly established by selecting the best \(\alpha \times 100\% \;(\alpha \in (0,1))\) of offspring in the current population, and two social knowledge matrices are subsequently obtained by extracting the scheduling and distribution knowledge of \(\Omega_{S}\) as follows:

$$ SKP = \left[ {\begin{array}{*{20}l} {\begin{array}{*{20}c} {J_{11} \;} & {\;\;J_{12} } & {\; \cdots } & {J_{1j} } & {\; \cdots } & {\;J_{1n} } \\ \end{array} } \hfill \\ {\begin{array}{*{20}c} {J_{21} \;} & {\;\;J_{22} } & {\; \cdots } & {J_{2j} } & {\; \cdots } & {J_{2n} } \\ \end{array} } \hfill \\ {\;\; \vdots \;\;\;\;\;\;\;\;\;\; \vdots \;\;\;\;\;\;\;\;\;\;\;\;\;\; \vdots \;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\; \vdots \;} \hfill \\ {\begin{array}{*{20}c} {J_{i1} \;} & {\;\;J_{i2} } & {\; \cdots } & {\;J_{ij} } & {\;\; \cdots } & {J_{in} } \\ \end{array} } \hfill \\ {\;\; \vdots \;\;\;\;\;\;\;\;\;\; \vdots \;\;\;\;\;\;\;\;\;\;\;\;\;\; \vdots \;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\; \vdots \;} \hfill \\ {\begin{array}{*{20}c} {J_{n1} } & {\;\;J_{n2} } & {\; \cdots } & {J_{nj} } & {\;\; \cdots } & {J_{nn} } \\ \end{array} } \hfill \\ \end{array} } \right],1 \le i,j \le n $$
(12)
$$ SKB = \left[ {\begin{array}{*{20}l} {\begin{array}{*{20}c} {B_{11} \;} & {\;\;B_{12} } & {\; \cdots } & {\;B_{1b} } & {\; \cdots } & {\;B_{1n} } \\ \end{array} } \hfill \\ {\begin{array}{*{20}c} {B_{21} \;} & {\;\;B_{22} } & {\; \cdots } & {\;B_{2b} } & {\; \cdots } & {B_{2n} } \\ \end{array} } \hfill \\ {\;\; \vdots \;\;\;\;\;\;\;\;\;\;\; \vdots \;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\; \vdots \;\;\;\;\;\;\;\;\;\;\;\;\;\;\; \vdots \;} \hfill \\ {\begin{array}{*{20}c} {B_{j1} \;} & {\;\;B_{j2} } & \cdots & {\;B_{jb} } & {\; \cdots } & {\;B_{jn} } \\ \end{array} } \hfill \\ {\;\; \vdots \;\;\;\;\;\;\;\;\;\;\; \vdots \;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\; \vdots \;\;\;\;\;\;\;\;\;\;\;\;\;\;\; \vdots \;} \hfill \\ {\begin{array}{*{20}c} {B_{n1} } & {\;\;\;B_{n2} } & {\; \cdots } & {\;B_{nb} } & {\; \cdots } & {\;B_{nn} } \\ \end{array} } \hfill \\ \end{array} } \right],1 \le j,b \le n $$
(13)

where \(J_{ij}\) denotes the number of times that job \(j\) appears on position \(i\) in the job sequence, \(B_{jb}\) represents the number of times that job \(j\) is allocated to batch \(b\), and n is the number of jobs released to distributed HFSs. Therefore, the matrices of \(SKP\) and \(SKB\) reserve the knowledge of job processing and batch assignment, respectively.

According to the above two matrices extracted from \(\Omega_{s}\), an offspring individual performs social learning as Eqs. (14) and (15) to generate better production and distribution schedules. In these two equations, \(\Omega_{u}^{i}\) is the set of unassigned jobs until the ith column of an individual, and \(x_{i}\) and \(y_{i}\) indicate the job and its assigned batch in the ith column, which are determined by the job with the largest \(J_{ij}\) in \(SKP\) and the batch with the largest \(B_{jb}\) in \(SKB\) respectively.

$$ x_{i} = \arg \;\mathop {\max }\limits_{{j \in \Omega_{u}^{i} }} J_{ij} ,1 \le i,j \le n $$
(14)
$$ y_{i} = \arg \;\mathop {\max }\limits_{b = 1,2, \ldots ,n} B_{{x_{i} b}} ,\quad 1 \le i,b \le n $$
(15)
  • Family learning operation

As one important form of inter-generational learning, family learning has a major impact on personal development throughout life. Therefore, it is very common that a person uses the knowledge and experience of their ancestors to avoid mistakes and improve decision-making practice. To mimic the family learning process, each of the offspring individual employs two matrices of FKP and FKB, which have the same structures as SKP and SKB, to store its family knowledge on scheduling and distribution. Different from the social learning operation, FKS and FKD of an offspring individual are established on a set of its ancestors rather than an elite social set \(\Omega_{S}\) of current population. Similar to social learning, the offspring individuals improves their performance using the knowledge of job processing and batch assignment stored in FKS and FKD.

  • Individual random learning operation

A person tends to conduct a random learning for new problem solving, especially in the absence of priori knowledge and experience. To imitate the randomness of human learning, an individual random learning operation is adopted to update an offspring as follows:

$$ x_{i} = randsample\left( {\Omega_{u}^{i} } \right),\quad 1 \le i \le n $$
(16)
$$ y_{i} = randint\left( {1,\;n} \right),\quad 1 \le i \le n $$
(17)

where \(randsample(\Omega_{u}^{i} )\) is a function that returns a random job from the unsigned job set \(\Omega_{u}^{i}\), \(randint\left( {min,max} \right)\) is a function that returns an integer uniformly distributed in the range of min and max.

To improve the offspring generated by crossover and mutation, each of the genes in a chromosome performs one type of learning. Based on two predefined parameters \(P_{1}\) and \(P_{2}\), the selection probability of social learning, family learning, and individual random learning is determined by the values of \(P_{1}\), \(P_{2} - P_{1}\), and \(1 - P_{2}\).

Since \(P_{1}\) and \(P_{2}\) greatly affect the learning performance of offspring individuals, an adaptive learning scheme is applied as Eqs. (18) and (19) to adjust these two parameters throughout the evolution process.

$$ P_{1} = P_{1}^{min} + \frac{{\left( {P_{1}^{max} - P_{1}^{min} } \right)T_{cur} }}{{T_{max} }} $$
(18)
$$ P_{2} = P_{2}^{max} - \frac{{\left( {P_{2}^{max} - P_{2}^{min} } \right)T_{cur} }}{{T_{max} }} $$
(19)

where \(P_{{_{1} }}^{min}\) and \(P_{{_{1} }}^{max}\) indicate the minimum and maximum values of \(P_{1}\), \(P_{2}^{min}\) and \(P_{2}^{max}\) represent the minimum and maximum values of \(P_{2}\), and \(T_{cur}\) and \(T_{max}\) are the current time and the maximum computation time of AHLBGA respectively. According to the above two equations, the adaptive learning scheme allows more individual random learning along with population evolution to avoid getting trapped into local optimum, and eventually generates potentially better solutions.

Based on the above details of adaptive learning operator, the complete learning process of an offspring individual is summarised as follows:

figure d

4 Experimental results and discussion

To validate the effectiveness of proposed heuristics and AHLBGA for addressing the IPDSP in distributed HFSs, this section compares them with some competitive algorithms. All the algorithms were developed using C++ in CodeBlocks 16.01 and ran on a PC with Intel® Core™ i5-5200U 2.20 GHz CPU and 4 GB RAM under Microsoft Windows 10 environment. The following subsections present the experiment design, parameter tuning, and experiment results of the compared algorithms, respectively.

4.1 Experiment design

The IPDSP with factory eligibility and 3PL distribution is a new combinatorial scheduling problem and therefore no benchmark instances have been found in the literature. To evaluate the performance of proposed heuristics and AHLBGA, two sets of test problems are randomly generated, namely small-sized problems with f = {2, 3, 4, 5}, n = {20, 50}, t = {5, 10}and large-sized problems with f = {2, 3, 4, 5}, n = {100, 200}, t = {15, 20}. For each of the test problems, the number of parallel machines in a stage is randomly selected in the range [2, 5]. The time-related and cost-related parameters of test problems are shown in Table 1. In this table, Unidrnd(a, b) returns a uniform random integer in the interval [a, b].

Table 1 Time-related and cost-related parameters

As a new combinatorial optimization problem, there are no scheduling algorithms specifically designed for the IPDSP with factory eligibility and 3PL distribution. Instead of using some existing algorithms, the proposed three fast heuristics (CR-based, EDD-based, and SLACK-based heuristics) and AHLBGA are compared with the following two competitive algorithms:

  • GA: As a standard GA, this algorithm does not employ the proposed learning operator to improve the offspring individuals. It starts with a random initial population, and adopts the same selection, crossover and mutation operators as used in the proposed AHLBGA for reproduction.

  • Human-learning-based GA (HLBGA): This algorithm is the same as AHLBGA, except for determining the selection probability of learning activities using \(P_{1} = 1/3\) and \(P_{2} = 2/3\) throughout the evolution process.

As a commonly used performance measure, the relative error (RE) is applied to evaluate the effectiveness of scheduling algorithm as follows:

$$ RE = \frac{{C_{i} - C_{min} }}{{C_{min} }} \times 100 $$
(20)

where \(C_{min}\) represents the minimum total cost of earliness, tardiness and delivery obtained by all the compared algorithms, and \(C_{i}\) indicates the total cost of earliness, tardiness and delivery obtained by a specific algorithm i.

Among all the compared algorithms, RE is directly used to evaluate the performance of CR-based, EDD-based, and SLACK-based heuristics. To obtain more reliable experimental results of GA, HLBGA, and AHLBGA, we ran these three meta-heuristics 20 independent times for each of the test problems and measured their performance by the best RE (BRE), the average RE (ARE), and the worst RE (WRE). Accordingly, a meta-heuristic with lower BRE, ARE, and WRE is capable of generating better integrated production and distribution schedules. For all the compared meta-heuristics, the maximum computation time required to solve each of the test problems is set to 0.1 × n × t (s).

4.2 Parameter tuning with Taguchi method

The appropriate parameter setting plays a significant role in the effectiveness of meta-heuristics. The performance of AHLBGA greatly depends on four key parameters, namely population size (\(PS\)), crossover rate (\(P_{c}\)), mutation rate (\(P_{m}\)), and the population ratio to establish the social knowledge matrices (\(\alpha\)). The Taguchi method provides a systematic and efficient way to determine near optimum design parameters. Compared with a full factorial design, this method can significantly reduce the number of required experiments by using an orthogonal array and is therefore adopted to tune the parameters of AHLBGA.

The Taguchi method is implemented on a moderate-sized IPDSP with \(f\) = 4, \(n\) = 50 and \(t\) = 10. Table 2 describes the key parameters and their levels considered in the Taguchi method. Based on an orthogonal array L16 (44), different parameter combinations shown in Table 3 are examined. AHLBGA with a specific parameter combination firstly runs 10 times to address the selected IPDSP, and ARE is subsequently obtained as the responsible variable (RV) value. Instead of conducting 44 = 256 experiments in the full factorial design, the Taguchi method only performs 16 experiments to determine the parameters of AHLBGA.

Table 2 Factors and their levels of AHLBGA
Table 3 Combinations of different factor levels using an Orthogonal array L16 (44)

According to the ARE of Table 3, Table 4 ranks the four factors based on their Delta statistics, each of which is the difference of maximum and minimum values of average ARE for a specific factor. From the ranks in this table, it is clear that \(P_{c}\) has the most significant effect on the performance of AHLBGA.

Table 4 Average ARE at each factor level

Based on the results of Table 4, the trend of each parameter is depicted in Fig. 7. Since the studied IPDSP is a minimization type problem, AHLBGA tends to perform better when these parameters are set to the values with the minimum average ARE. A larger value of \(PS\) allows to explore more solutions of the search space and the value of \(PS\) is set to 200 according to Fig. 7. In addition, a moderate is capable of maintaining a better balance between population diversity and convergence performance, and therefore the best 20% of offspring individuals are chosen to extract social knowledge in this study. Furthermore, the values of and can be easily determined by Fig. 7. From the above analysis, a good choice of parameter setting is suggested as,,, and. For a fair performance comparison among the scheduling algorithms,,, and of GA and HLBGA are set to the same values as used in AHLBGA.

Fig. 7
figure 7

Factor level trend of AHLBGA

4.3 Experiment results and discussion

To evaluate the performance of the proposed fast heuristics and AHLBGA, both small-sized and large-sized test problems shown in Tables 5 and 6 are tested. In these two tables, f, n, and t indicate the numbers of factories, jobs, and stages in the distributed FFS environment, respectively. For each of the compared meta-heuristics, the maximum running time is set to 0.1 × n × t (s) in this study and the CPU time is therefore in the range between 10 and 400 s. Compared with the meta-heuristics, the three heuristics are much faster and give a solution within less than 1.9 s. For better performance comparison, the best BRE, ARE, and WRE for each of the test problems in these two tables are highlighted in bold.

Table 5 Computation results of SLACK, EDD, CR, GA, HLBGA, and AHLBGA for small-sized problems
Table 6 Computation results of SLACK, EDD, CR, GA, HLBGA, and AHLBGA for large-sized problems

According to the experiment results on the two sets of test problems, the following are observed:

  1. 1.

    Among of the six scheduling algorithms, ARE of GA, HLBGA, and AHLBGA is much less than RE of the three fast heuristics. The better performance of the three meta-heuristics results from the stochastic search of solution space, which is capable of offering a good balance between exploration and exploitation.

  2. 2.

    Compared with the SLACK and EDD heuristics, the CR heuristic provides better results on average, indicating that the CR rule is more effective to reduce the total cost of earliness, tardiness and delivery.

  3. 3.

    Since HLBGA and AHLBGA further improve the offspring individuals by a learning operator after crossover and mutation, they perform better than GA in terms of BRE, ARE, and WRE over all the test problems. Furthermore, AHLBGA outperforms HLBGA when generating integrated production and distribution schedules. The superiority of AHLBGA lies in the effectiveness of the adaptive learning operator to guide the learning process of offspring individuals.

Based on the experiment results of the above two tables, the analysis of variance (ANOVA) and the least significant difference (LSD) test are conducted to validate whether the performance differences among the compared algorithms are statistically significant. With regard to small-sized and large-sized test problems, Figs. 8 and 9 show the means plots with LSD intervals (95% confidence level), respectively. As shown in these two figures, AHLBGA provides the statistically best performance when generating the integrated production and distribution schedules in distributed HFS environments.

Fig. 8
figure 8

Means plot and LSD intervals of different algorithms for small-sized problems

Fig. 9
figure 9

Means plot and LSD intervals of different algorithms for large-sized problems

To further evaluate the performance of SLACK, EDD, CR, GA, HLBGA, and AHLBGA, the trends of ARE for different sizes of f, n, and t are presented in Figs. 10, 11 and 12, respectively. From these figures, AHLBGA provides the least ARE with slight fluctuation for all sizes of f, n, and t, indicating its effectiveness and robustness for solving the IPDSP with factory eligibility and 3PL distribution. As described in Fig. 10, ARE obtained using these six algorithms is not significantly changed, implying that the number of factories (f) has little influence on the performance of all the compared algorithms. In addition, the overall ARE values of the six algorithms in Figs. 11 and 12 show an increasing trend when addressing large-sized test problems (n = 100, 200; t = 15, 20).

Fig. 10
figure 10

ARE with different numbers of factories (f)

Fig. 11
figure 11

ARE with different numbers of jobs (n)

Fig. 12
figure 12

ARE with different stages in a factory (t)

5 Conclusion

This paper focuses on a new integrated production and distribution scheduling problem (IPDSP), in which jobs are firstly processed in distributed hybrid flow shops (HFS) and subsequently delivered to a customer in batches. To satisfy the industrial practice of production and distribution, both factory eligibility and multiple transportation modes are considered. Accordingly, in the studied IPDSP, some products can only be manufactured in a subset of distributed factories, and the transportation of finished products is conducted by the manufacturer or the 3PL provider.

To minimize the sum of earliness, tardiness and delivery costs, three fast heuristics (CR-based, SLACK-based, and EDD-based heuristics) and an adaptive human-learning-based genetic algorithm (AHLBGA) are developed to generate integrated production and distribution schedules. These three heuristics employ simple scheduling rules to sequence the released jobs and evenly group them into batches for production and distribution. Different from the traditional GA, AHLBGA hybridizes an adaptive learning operator with crossover and mutation to enhance the global and local search ability. Motivated by human learning behaviours, the learning operator involves social learning, family learning, and individual random learning operations. To provide a better trade-off between search exploration and exploitation, the selection probability of a learning operation is dynamically adjusted throughout the evolution process. To evaluate the effectiveness of proposed algorithms, two sets of randomly generated IPDSPs are established. Experimental results indicate that AHLBGA offers the best performance among all the compared scheduling algorithms. The superiority of AHLBGA results from incorporating three types of human learning operations into the evolution process of GA. Accordingly, AHLBGA has the potential to provide MTO manufacturers with better operational decisions in distributed HFS environments.

Future research may focus on integrated production and distribution scheduling in more realistic industrial environments. It would be interesting to extend the proposed algorithms to solve the IPDSP with batch delivery to multiple customers at different locations, such that vehicle routing decisions have to be made. Since inventory buffers exist between the production and distribution stages, another interesting extension of IPDSP is to take into account inventory characteristics, such as holding cost and inventory capacity. Furthermore, apart from the proposed three simple learning operations of AHLBGA, more sophisticated human learning mechanisms may be developed to enhance the search ability of GA.