1 Introduction

The railway has always been considered by the people and supported by governments as an efficient transportation system. In addition to critical features such as reliability, comfort, and safety, its essential feature of economic fuel consumption, which also significantly reduces greenhouse gas emissions, has made it a green transportation system. Due to the increasing demand for travel and freight by rail network, having efficient computer-based models in response to challenges and complexities such as train scheduling or rescheduling, passenger waiting time, and arrival time is inevitable.

Improving scheduling and providing passenger satisfaction with increasing travel demand in different seasons conflicts with the available resource limitation. On the one hand, increasing the number of trains leads to decreasing the waiting time and increasing passenger satisfaction. In contrast, operating costs have increased, and traffic management, locomotive, and crew schedules have become more complex. Train schedule depends on many conflicting factors and parameters, such as resource constrains, continuous parameter changes, and various passenger requests. Different models, such as MILP, are utilized to solve these conflicts. [1, 2]. Train scheduling computation usually requires several hours or even a few days of utilizing computational units, high-performance hardware resources, and high-processing units. Due to contingency rescheduling and time constrains, train scheduling with meta-heuristic algorithms is more effective [3].

NSGA-II is generally utilized to solve various scientific problems such as non-classified multi-objective optimization or multi-dominance characteristics with five features, fast crowded distance, well diversity, well convergence, fast non-dominated rank, and exclusive observant approach [4]. As a stochastic algorithm, NSGA-II comprises six steps: initialization, sorting, crowding distance, selection, genetic operators, and recombination to transfer the population to the next generation. The population is processed and handed over to the next step at each step to get enough maturity for the final mutation [5]. Despite its numerous advantages, NSGA-II suffers from some drawbacks, such as a non-dominated sorting process on 2N size, the probability of convergence constrain due to crowded comparison, the possibility of restrictions on the uniform distribution of space in some problems, and the possibility of producing a similar population. Creating a new generation makes identification much more complicated [6, 7].

The SA algorithm is a meta-heuristic method that provides near-optimal solutions for big combined optimization problems by sampling high-temperature solids and gradually decreasing the temperature [8]. In searching the solution space, SA tries not to get stuck in the trap of local optimum points with a neighborhood function and to find new solutions close to global optimum points [9]. The SA algorithm in search of optimal points generates and evaluates a random point in the neighborhood of the current response and the specified temperature T. Repeating the process of reducing T and searching for new points continues until the desired result is achieved. The probability of accepting new searched answers will be generous at high temperatures and greedy as the temperature decreases. The advantages of the SA algorithm are the possibility of theoretical proof and ease of implementation with appropriate efficiency [10]. Despite the numerous advantages of the SA algorithm, this algorithm also has some disadvantages, including the low convergence speed [11,12,13].

Based on the abovementioned discussion, each NSGA-II and SA algorithm has some disadvantages. To tackle these disadvantages, the mentioned algorithms are combined as a hybrid form to solve the problem, and to utilize a random guided hybrid, search method, a homogeneous set of more optimal solutions is created on the Pareto front [14, 15]. Each chromosome of the population calculates its value based on a fitness function, and using mutation and crossover actions, new solutions are improved [16]. One of the advantages of hybrid optimization is finding new solutions that each algorithm alone cannot find near the global optimum. Also, creating a balance between exploration and exploitation in the best possible way is another advantage of hybrid optimization in terms of accuracy and execution time [17, 18].

Multi-object Evolutionary Algorithm with Hybrid Sampling Strategy (MOEA-HSS) is a multi-objective meta-heuristic algorithm with a hybrid elitist sampling strategy that takes selective solutions using the vector genetic method from the edge vector evaluated genetic algorithm (VEGA) and central area Pareto dominating and dominated relationship-based fitness function (PDDR_FF) of Pareto front. Sampling for each objective function is done at the maximum size of half of the original population, utterly independent of other objective functions, and based on a proportionality function. They converge toward the central area of the edge of the Pareto front [4, 6].

Multi-object Evolutionary Algorithm Based on Decomposition (MOEA/D) has a pleasing efficiency in solving the multi-objective optimization problem. This algorithm is more effective than other evolutionary algorithms in the dominant Pareto space for solving a large optimization problem. One of the advantages of decomposing the multi-objective optimization problem into a few simple scalar optimization problems and computational complexity is less than evolutionary algorithms. However, one of the disadvantages of this algorithm is the low speed of convergence and the lack of population diversity compared to other evolutionary algorithms [19, 20].

A new metaheuristic algorithm is recently being developed and gained attention in the research community. The Multi-Objective Grey Wolf Optimizer (MOGWO) is an algorithm for optimization problems with multiple objectives [21]. The MOGWO is the developed version of the Grey Wolf Optimizer (GWO) algorithm [22]. The algorithm simulates the leadership hierarchic structure and hunting manner of grey wolves in nature. Four types of grey wolves, namely alpha, beta, delta, and omega, are utilized for simulating the leadership hierarchical structure. The optimization in three steps is implemented including search hunting, encircling, and attacking prey [23].

A fixed-size archive is integrated with the GWO algorithm for storing and retrieving non-dominated solutions. The optimal responses are compared with the individuals' archives in the iteration process, and then, the non-dominant answer is replaced by a random individual from the archives [23]. The MOGWO algorithm can present very competitive results compared to other meta-heuristics algorithms for solving multi-objective problems [21, 24].

The NSGA-II and SA are still widely studied and used in the literature on optimization problems because they have different strengths and limitations. NSGA-II has an excellent performance at exploring the search space and finding solutions quickly [25]. On the other hand, SA is good at escaping the trap of local optimum [8]. Hybridization of NSGA-II and SA is more efficient and effective by overcoming their limitations, combining their strengths, and using them simultaneously to solve optimization problems [16]. This hybridization has various advantages including:

  1. 1.

    Improved global optimization: Both NSGA-II and SA algorithms are performing well in finding global optima. Simultaneous use of both algorithms increases the chance of global optimization.

  2. 2.

    Enhanced exploration and exploitation: NSGA-II is good at exploring the search space, while SA is excellent at exploiting the search space hybridization can lead to a better balance between exploration and exploitation.

  3. 3.

    Faster convergence: NSGA-II is known to converge slowly toward the optimal solution, while SA converges faster. The hybrid approach has faster convergence and better performance.

  4. 4.

    NSGA-II and SA can be customized for specific optimization problems [26,27,28].

In this article, to improve the quality of the answers to the multi-objective optimization problem of train schedules, the NSGA-II algorithm is hybrid with the SA operator as the main framework, and multi-crossover and multi-mutation methods are used to produce a faster and better next-generation. To prove the superiority of NSGA-II-SA solutions, three other multi-objective optimization algorithms, namely NSGA-II, MOEA-HSS, and MOEA/D, are implemented, and an experiment and Pareto front analysis is performed. In another approach, by selecting one solution from NSGA-II-SA, a comparison will be made with the real train scheduling data.

This study's contributions are listed as the following:

  • A developed mathematical scheduling model is proposed to reduce the waiting time of passengers and the number of train scheduling cycles under conditions of uncertainty.

  • The meta-heuristic NSGA_II algorithm is a hybrid with an SA operator, which has a good convergence speed, suitable variety, and is entirely uniform.

  • Four algorithms solve the proposed model, and the Pareto front solutions are analyzed with several metrics.

  • In a case study, using the decomposition approach and selecting the gold point, the superiority of the hybrid NSGA-II-SA algorithm response is proven in terms of train scheduling with real data.

The rest of this article is written as follows. Section 2 refers to recently published papers in train scheduling and meta-heuristic multi-objective algorithms. Section 3 refers to the presentation of a mathematical model to formulate a multi-objective function for train scheduling. Section 4 refers to the detailed description of the hybrid NSGA-II algorithm with the SA operator for the proposed model. Section 5 discusses the numerical experiments, and the comparison of the results of four algorithms in the Pareto front space and the comparison of the results of the NSGA-II-SA algorithm are compared with the real values of railway scheduling. Finally, Sect. 7 refers to conclusions and suggestions for future studies in train scheduling by meta-heuristic algorithms.

2 Literature overview

Serval studies in research and academic centers regarding improving train scheduling, passenger satisfaction, and reducing operational costs are being carried out. This section examines the optimization problems articles and the use of meta-heuristic algorithms in transportation, especially rail, in recent years.

The paper [29] presented a multi-objective optimization of train schedules, passenger traffic control, and stop patterns. A mixed integer nonlinear programming (MINLP) model is applied to balance operating costs and service level performance. The robust optimization model is used to balance solution robustness and scenario-based model robustness. Finally, this research demonstrates the accuracy of the proposed model with numerical experiments according to the Beijing metro line data set.

The paper [30] investigated train scheduling problems considering energy consumption efficiency and operation cycle on urban railways. The aim is to solve the problem of the train operating cycle, the use of regenerative braking energy, and the energy saving of the traction motor. This paper utilized a particle swarm algorithm for optimizing train scheduling and operating cycle. Finally, the numerical experiments indicate reduced operating costs compared with the initial train scheduling on Guangzhou Subway Line 9.

The paper [31] presented a novel approach to optimizing train scheduling and rolling stock operating cycles based on dynamic passenger demand and transport capacity. Two integer linear programming models have optimized for total cost and time–space by flexible coupling/splitting activities. A diving heuristic algorithm optimizes train scheduling and rolling stock operating cycle. Finally, this research demonstrates the efficiency of the proposed methods and algorithm with numerical experiments compatible with the Shanghai Subway line real data set.

The paper [32] the train timetable developed based on passenger transport priority and minimum cargo transfer delay using an objective function. An analysis of multi-modal scheduling solutions and demand setting investigated in a simulation. The results demonstrate that obtained the train schedule increases service quality and reduces freight demand.

The paper [33] developed an optimization multi-module model by utilizing the accessible route in the primary train schedule. This model solves a repeatable framework of the Lagrangian relaxation by the branch-and-bound algorithm. It aims to generate the train schedule by resolving constrains and daily operational conflicts. Finally, this paper shows the validity of the proposed approach with the INFORMS RAS 2016 (Railway Applications Section) real dataset.

The paper [34] investigated multi-objective optimization problems of train operation with station classification and passenger arrival rate. The multi-layer programming model is provided with train stop optimization to minimize operation costs and travel costs. The proposed model is implemented with a meta-heuristic SA algorithm and sequential average method. Finally, this research discusses the effectiveness of the results using sensitivity analysis on the operation cost and the travel cost reduction with actual data of Shanghai Metro.

The paper [35] investigated the problem of optimizing train scheduling based on dynamic passenger demand on the double-track railway in Turkey, and a linear programming model has been formulated to minimize the average passenger waiting time. The model investigates resource limitations, train capacity, and fleet size of passengers remaining from the first train. Finally, this research discusses the sensitivity analysis between passenger satisfaction and operating costs by comparing railway traffic scheduling.

The paper [1] presented a multi-objective optimization of train scheduling and a robust stop schedule by uncertainty. The aim is to find in order robust solutions to improve passenger satisfaction under uncertainty, and a MILP model has been provided based on the light robustness technique for passengers in each pair of origin and destination stations for each train. Finally, this research discusses the effectiveness of the nominal passenger demand reduction with real-life data of the Wuhan–Guangzhou high-speed railway.

The paper [16] studied minimizing the total transportation costs, the allocation of resources, and risks of disruption in the hub of the Turkish postal transport network. This research combines meta-heuristic SA and Tabu Search (TS) [36] algorithms to improve the solution accuracy and computational time. The result is compared with other meta-heuristic algorithms.

The paper [6] presented a meta-heuristic algorithm by parallel processing with hybrid sampling and mutation-based learning for computation speed improvement. A multi-objective optimization model is solved with the MOEA-HSS and several meta-heuristic algorithms. The model aims to minimize the average passenger waiting time and reduce the average train circulation. This research shows that the MOEA-HSS algorithm needs less average execution time and outperforms the NSGA-II algorithm, and Power Pareto Evolutionary Algorithm (SPEA2) [37].

The paper [38] studied the scheduling problem of uncertain processing with a discrete scenario set. This research combines the NSGA-II algorithm and the SA algorithm to minimize the average execution time and the execution time of the worst under scenarios. The studied results show the developed algorithm performs better than the other four meta-heuristic algorithms defined in the paper. Table 1 includes the related works in optimization to transportation, problem model, model type, objective function, and problem-solving approach.

Table 1 The related works in the optimization to solve transport network and solution approach

3 Problem definition

In this section, mathematical formulas for the railways scheduling model, parameters, constrains, and variables in uncertainty conditions are presented [6]. In Fig. 1, a modeling diagram is illustrated for a double-track railways network, where the route has permission for a single direction. The network consists of 2S stations, with stations 1 to S for the departure route and stations S + 1 to 2S for the return path. Trains movement are designed not to overtake each other along the route or station, and a safe headway has guaranteed for two consecutive trains. The operating period P continues with trains’ movement from the first station until the final station S and then comes back from the S + 1 station to the 2S station. As a result, they complete an operation cycle c.

Fig. 1
figure 1

Schematic double-track railway network

Other assumptions are as follows: (1) The capacity of the train and the station is not taken into consideration; (2) no passenger is left behind at the train station; (3) only one train stops at each station simultaneously; and (4) scheduling is presented for one period. Table 2 describes the indicators, parameters, and decision variables of the proposed model.

Table 2 Indexes, parameters, and decision parameters

The objective functions (1) and (2) are defined to minimize the average passenger waiting time at the stations and the total number of train operation cycles during an active period. Increasing the number of trains operation can enhance passenger satisfaction but lead to higher train operating costs. Based on the above description, the objective functions and problem constrains are stated in constrains (3)–(14) as follows:

The objective function (1) corresponds to the minimized average waiting time of the passenger, \({p}_{ui}\) is a parameter including the probability for passenger arrival rate at station i under scenario u. The \({a}_{ui}\left(p\right)\) is a parameter showing the number of passengers arriving at station i during period p under scenario u. The \({t}_{ui}^{PA}\left(p\right)\) is another parameter to determine the passenger arrival time at Station i under scenario u. The \({t}_{ucki}^{D}\) is a decision variable that includes the arrival time \({\text{t}}_{\text{ucki}}^{\text{A}}\), stop time at the station \({\text{t}}_{\text{ucki}}^{\text{S}}\), and train turnaround time \({{t}_{u}}^{T}\). The turnaround time is only considered at the first and last stations.

$$ \min Z_{1} = \mathop \sum \limits_{{p = P_{start} }}^{{P_{end} }} \mathop \sum \limits_{u = 1}^{U} \mathop \sum \limits_{c = 1}^{C} \mathop \sum \limits_{k = 1}^{N} \mathop \sum \limits_{i = 1}^{2S} (p_{u} a_{ui} \left( p \right)(t_{ucki}^{D} - t_{ui}^{PA} \left( p \right)))/\mathop \sum \limits_{u = 1}^{U} \mathop \sum \limits_{i = 1}^{2S} p_{u} a_{ui} \left( p \right) $$
(1)

The objective function (2) corresponds to minimizing the total number of operating cycles during the operating period. The \({\text{x}}_{\text{ucki}}\left({\text{p}}\right)\) is a binary decision variable to control the possibility of passengers getting on the trains. If \({\text{x}}_{\text{ucki}}\left({\text{p}}\right)\) =1, it means that the passenger has enough time to get on the train.

$$ Min \, Z_{2} = \mathop \sum \limits_{{p = P_{start} }}^{{P_{end} }} \mathop \sum \limits_{u = 1}^{U} \mathop \sum \limits_{k = 1}^{N} \mathop \sum \limits_{i = 1}^{2S} p_{u} x_{uki} \left( p \right) $$
(2)

Constrain (3) describes the train's arrival time to station i + 1, which includes the departure time \({\text{t}}_{\text{ucki}}^{\text{D}}\) from the previous station i and the travel time \({\text{t}}_{\text{ucki}}^{\text{R}}\) between two stations i, and i + 1.

$$ {\text{t}}_{{\text{ucki + 1}}}^{{\text{A}}} {\text{ = t}}_{{{\text{ucki}}}}^{{\text{D}}} {\text{ + t}}_{{{\text{ucki}}}}^{{\text{R}}} { ,}\quad\forall {\text{ c, k,i }} \ge {\text{2 , u}} $$
(3)

Constrain (4) shows the departure time of the train at station i, which includes the arrival time \({\text{t}}_{\text{ucki}}^{\text{A}}\) and stopping time \({\text{t}}_{\text{ucki}}^{\text{S}}\) of the train at station i the turnaround time \({{\text{t}}_{\text{u}}}^{\text{T}}\) is only considered at the first and last stations.

$$ {\text{t}}_{{{\text{ucki}}}}^{{\text{D}}} {\text{ = t}}_{{{\text{ucki}}}}^{{\text{A}}} {\text{ + t}}_{{{\text{ucki}}}}^{{\text{S}}} {\text{ + t}}_{{\text{u}}}^{{\text{T}}} { ,}\quad\forall {\text{ u, c, k, i}} $$
(4)

Constrain (5) is defined to control of the train traveling time between station i and station i + 1. The \({\text{t}}_{\text{ui}}^{\text{Rmax}}\) and \({\text{t}}_{\text{ui}}^{\text{Rmin}}\) are parameters to control the maximum and minimum traveling time between stations i and i + 1.

$$ {\text{t}}_{{{\text{iu}}}}^{{{\text{Rmin}}}} \le {\text{t}}_{{{\text{ucki}}}}^{{\text{R}}} \le {\text{t}}_{{{\text{iu}}}}^{{{\text{Rmax}}}} { ,}\quad\forall {\text{ u, c, k, i}} $$
(5)

Constrains (6) and (7) describe safe headway time for two consecutive trains. The \(h_{u}^{\max }\) and \(h_{u}^{\min }\) are parameters for controlling the maximum and minimum headway between two trains.

$$ {\text{h}}_{u}^{{{\text{min}}}} \le {\text{t}}_{{\text{ck + 1,iu}}}^{{\text{D}}} -{\text{ t}}_{{{\text{ucki}}}}^{{\text{D}}} \le {\text{h}}_{u}^{{{\text{max}}}} { ,}\quad\forall {\text{ u, c, k, i}} $$
(6)
$$ {\text{h}}_{{\text{u}}}^{{{\text{min}}}} \le {\text{t}}_{{\text{uck + 1,i}}}^{{\text{A}}} - {\text{t}}_{{{\text{ucki}}}}^{{\text{A}}} \le {\text{h}}_{{\text{u}}}^{{{\text{max}}}} { , }\quad\forall {\text{ u, c, k, i}} $$
(7)

Constrain (8) describes passenger waiting time \({\text{t}}_{\text{ucki}}^{\text{W}}\) at the station. It must always be greater or equal to the train departure time. M is a parameter with a high positive value. When the train departure time \({\text{t}}_{\text{ucki}}^{\text{D}}\) is earlier than the passenger arrival time \({ t}_{i}^{PA}\left(p\right)\), in this case, the passenger waiting time \({\text{t}}_{\text{ucki}}^{\text{W}}\) will be zero.

$$ {\text{t}}_{{{\text{ucki}}}}^{{\text{D}}} - {\text{ t}}_{{{\text{iu}}}}^{{{\text{PA}}}} \left( {\text{p}} \right){ - }Mx_{ucki} \left( p \right) \le {\text{t}}_{{{\text{ucki}}}}^{{\text{W}}} { ,}\quad\forall {\text{ c, k, i, u, }}p \in { }\left[ {{\text{P}}_{{{\text{start}}}} {\text{, P}}_{{{\text{end}}}} } \right] $$
(8)

Constrain (9) controls the train turnaround at the station.

$$ {\text{t}}_{{\text{ck, S + 1}}}^{{\text{A}}} - {\text{ t}}_{{\text{ck + 1,S}}}^{{\text{D}}} \ge {0 , }\quad\forall {\text{ c, k}} $$
(9)

Constrains (10) and (11) are defined to impose the value of positive variables.

$$ {\text{t}}_{{{\text{ucki}}}}^{{\text{A}}} {\text{,\,t}}_{{{\text{ucki}}}}^{{\text{D}}} {\text{,\,t}}_{{{\text{ucki}}}}^{{\text{R}}} {\text{,\,t}}_{{{\text{ucki}}}}^{{\text{S}}} {\text{,\,t}}_{{{\text{ucki}}}}^{{\text{W}}} \ge {0, }\quad\forall {\text{ u, c, k, i}} $$
(10)
$$ {\text{x}}_{{{\text{ucki}}}} \left( {\text{p}} \right) \in \left\{ {0,1} \right\}{,}\quad\forall {\text{ c, k, i, p}} \in \left[ {{\text{P}}_{{{\text{start}}}} {\text{, P}}_{{{\text{end}}}} } \right] $$
(11)

Constrains (12) and (13) calculate the traveling time between stations and headway between two consecutive trains.

$$ t_{ucki}^{R} = t_{iu}^{Rmin} + \left( {t_{iu}^{Rmax} - t_{iu}^{Rmin } } \right),\quad\forall \;u, \, c, \, k \, ,i $$
(12)
$$ h_{c} = h_{u}^{min} + \left( {h_{u}^{max} - h_{u}^{min} } \right) ,\quad\forall \;u, \, c, \, k \, ,i = \{ 1, \, S + 1 \, , \, 2S + 1, \, \left( {C_{max} } \right)S + 1\} $$
(13)

In constrain (14), the PU is a parameter to represent the probability of the number of passengers entering stations in each season. It is calculated using the Poisson distribution with \(\lambda_{u}\) as the mean number of events.

$$ P_{u} \sim Poisson(\lambda_{u} )\;,\forall \;u = \{ spring, \, summer, \, fall, \, w{\text{int}} er\} $$
(14)

4 Proposed solution

In recent years, meta-heuristic algorithms and one fitness function have been given much attention in solving engineering optimization problems due to their impressive efficiency [42]. This section describes the steps of developing the classic NSGA-II algorithm using multiple crossover and mutation operators, then the hybrid of NSGA-II with the SA operator. Finally, the proposed train scheduling model has been implemented using the developed hybrid algorithm.

4.1 Chromosome and population initialization

In algorithm 1, a random initial population is created and initialized based on input variables, namely the number of stations S, train cycles C, and scenario U (Fig. 2). First, the individual structure is created based on the problem's parameters and index (line 1). Then, the population matrix of the problem is produced in the required number based on U, C, and S (line 2). The main body includes three nested loops that initialize the population matrix (lines 3–5). Each individual will randomly have a value of 0 or 1 in the population matrix (line 6) [6].

Fig. 2
figure 2

Chromosome representation

figure a

4.2 Crossover process

The crossover operator is the critical leading force in finding new and optimal solutions and offers excellent solutions sometimes. Uniform and reduced surrogate operators are simultaneously utilized to improve the creation of the next generation [43].

4.2.1 Uniform crossover operator

As illustrated in Fig. 3, the classical crossover operator usually applies to chromosomes' predetermined location (s) in one or two fixed positions. In algorithm 2, the uniform operator does not need to determine the intersection point(s) in advance. The crossover process extends equally to the entire chromosome length. In this method, one binary string is randomly created with the same size as the chromosome (line 3). The exchange process of parents' genes will be performed to produce two offspring depending on a binary string value (lines 6, 7) [43].

Fig. 3
figure 3

Uniform crossover operator

figure b

4.2.2 Reduced surrogate crossover operator

In Fig. 4, the crossover operator may create a poor-quality population or analogous in some situations. The reduced surrogate operator will likely improve the production of the poor-quality population to an acceptable extent.

Fig. 4
figure 4

Reduced surrogate crossover operator

In algorithm 3, operations are avoided on chromosomes with similar genes. As a result, it will reduce the number of unnecessary executions of the fitness function and execution time. The parents' genes have been compared to a similar position, and differences are stored in a string (line 3). If the parents have at least one different gene, one place will be chosen randomly (lines 4, 5). The exchanging of genes will be done to produce two new offspring in the specified position (lines 6) [43].

figure c

4.2.3 Multi-crossover operator main body

In algorithm 4, the multi-crossover operators’ routine is provided as the replace surrogate and uniform operator. The crossover operator can be applied to a population or a percentage of the entire population according to the problem condition (lines 2, 3). The two parents have been randomly selected among the population (line 4). The reduced surrogate and uniform crossover routines are executed according to the value of a random binary number. This approach executes routines randomly and equally (lines 5–9).

figure d

4.3 Mutation process

The main effect of the mutation operator compared to the crossover operator is to preserve the diversity of the population and increase the probability of being close to the global optimum [43]. The probability of producing a diverse and elite population increases with the simultaneous use of several mutation operators with various characteristics. In the following, two replacement and insertion operators are presented to improve the production efficiency of the new generation [44].

4.3.1 Replacement mutation operator

In algorithm 5, the independent random replacement operator is presented to increase the offspring's diversity among the parents' population [44]. A gene from the parent is randomly selected during the replacement process (line 2). The operation of producing new offspring will be done by changing the value of the gene (line 3) (Fig. 5).

Fig. 5
figure 5

Replacement mutation operator

figure e

4.3.2 Insertion mutation operator

In algorithm 6, the insertion operator changes one or more genes from one chromosome to increase the ability to produce a new generation of offspring (Fig. 6). In the insertion process, several changeable genes are randomly selected from the parent chromosome based on the mutation rate and the size of the genes [44]. In the initialization, one offspring is placed equal to one parent (line 1). The number of parent genes is calculated to create the insertion process loop (lines 2, 3). A new offspring is created by randomly changing the gene position (Lines 4–5).

Fig. 6
figure 6

Insertion mutation operator

figure f

4.3.3 Multi-mutation main body operator

In algorithm 7, a multi-mutation method with replacement and insertion operator is presented to improve mutation efficiency in offspring production. An empty structure is created to store the new neighbors (line 1). The offspring is produced according to the ratio of the population and the percent probability of mutation (lines 2, 3). The mutation operator is performed only on one individual (line 4). The reduced surrogate and uniform crossover routines are executed by a binary variable random and equally (lines 6–9).

figure g

4.4 Simulated annealing (SA) process

Algorithm 8 describes the competition between the parent population and the candidate neighbors. A chromosome of the candidate neighbors competes with one chromosome of the parent population based on the SA rule. The number of iterations of the main loop is equal to the size of the population (line 1). The chromosome of the candidate neighbors is selected with more competency than the parent chromosomes by competition (line 2), and the neighbor replaces the parent (line 3). Otherwise, the ∆f is calculated as the difference between the value of the function neighbor and the parent (line 5). The candidate neighbor has a chance to replace the parent; if the value of \(e^{{{\raise0.7ex\hbox{${ - \Delta f}$} \!\mathord{\left/ {\vphantom {{ - \Delta f} T}}\right.\kern-0pt} \!\lower0.7ex\hbox{$T$}}}}\) is more prominent than a random number in the range {0, 1} (lines 6, 7), the candidate neighbor is selected and replaces the parent (line 8).

figure h

4.5 Evaluated process for solving the railway scheduling

Figure 7 shows the train movement graph schematically with decision variables and problem parameters for traffic management and control. It is the primary management tool in the railway dispatch centers for train planning, rolling stock movement, train rescheduling accident management, and all event control. It can also help better understand algorithm 9.

Fig. 7
figure 7

An overview of the train schedule graph

In algorithm 9, the train timetable and functions Z1 and Z2 create for individuals under four scenarios. This algorithm presents train scheduling on each station for each chromosome by the limitations of constrains (3)–(14). The algorithm starts with four main nested loops by the number of scenarios, operating cycles, several trains, and stations (lines 1–4). The train travel time \({t}_{ucki}^{R}\) is calculated as the difference between max and min travel time and the train travel time between two stations Si, and Si+1 (line 5 and used parameters \({t }_{ui}^{Rmax}, {t}_{ui}^{Rmin })\) ). The headway turnaround time \({h}_{c}\) has been calculated as the difference between the maximum and minimum headway parameters \({h}_{u}^{ max}- {h}_{u}^{min}\) and minimum headway \({h}_{u}^{min}\) between two consecutive trains (line 6). Suppose the train needs a turnaround, the train departure time \({t}_{ckiu}^{D}\) is replaced with headway turnaround time \({h}_{c}\) and train arrival time \({t}_{c-1.k.2Su}^{A}\) at the last station (lines 7, 8). Otherwise, the train arrival \({t}_{ck.i+1.u}^{A}\) and departure times \({t}_{ckiu}^{D}\) are calculated at two consecutive stations Si, and Si+1 (lines 10, 11). The train arrival time \({t}_{ucki}^{A}\) and departure time \({t}_{ucki}^{D}\) must have the minimum headway \({h}_{u}^{min}\) between two consecutive trains to guarantee safety essentials (lines 13–16). If the passenger arrival time parameter \({t}_{ui}^{PA}\left(p\right)\) is longer than the train departure time \({t}_{ucki}^{D}\) then calculate the value of function Z1 (line 18–20 and used parameters\({a}_{ui}\left(p\right) , {p}_{u},M\)). Calculation of the total average passenger waiting time is done in three steps. Firstly, the passengers' waiting times are calculated in line 21, secondly, the total number of passengers is calculated in line 29, and the value of the total average passengers' waiting times (Z1) is calculated in line 32.

figure i

4.6 Hybrid NSGA-II and SA generate Pareto solution and railway timetable

Figure 8 illustrates the steps of the hybrid NSGA-II algorithm and SA operator, which provides a big schematic view for a good understanding of algorithm 10.

Fig. 8
figure 8

Hybrid NSGA-II and SA

In algorithm 10, NSGA-II-SA will be initialized by the railway's data set, variables, and parameters in (3)–(14) (number of population, iteration, sub-iteration, temperature, Alpha …) (lines 1–3). The initial population is randomly created (line 4) and is evaluated by the fitness function (line 5). The non-dominated sorting and calculating crowding distance will be done on population by NSGA-II rule (lines 6–7, 16–17). The sorting operation will be done based on crowding distance; this approach ensures sorting the initial population based on ranking [11] (line 8, 18). With nested loops, the hybridization of NSGA-II and SA has been utilized to produce the next generation based on NSGA-II rules. The dominant generation is chosen with competition and more competency based on SA rules [45] (lines 9, 10). The multi-crossover and multi-mutation operators are utilized to generate for evaluating the population by the fitness function (lines 11–14). Offspring population S'(t) has merged with population multi-crossover Q(t) and multi-mutation R(t) (line 15). Superior neighbors S' (t) are inserted into S"(t) the size of the parent population P(t) (line 19).

The elite generation has been chosen among the neighbors S''(t) and the parent population P(t) by the SA rule (line 20) [38]. The dominant population will be stored as the best solution in the Pareto front (line 21). The temperature reduction is continued with a fixed and predetermined (Like Alpha) coefficient until the final result is obtained (line 22). Finally, the timetable has been extracted from the Pareto front population for the subsequent analysis (lines 26).

figure j

5 Discussion

In this section, the proposed model has been implemented by numerical calculation and evaluated and validated.

This model is implemented and executed with the developed NSGA-II-SA algorithm and four algorithms NSGA-II, MOEA-HSS, MOEA/D, and MOGWO.

Several measures are discussed to evaluate the efficiency of the algorithms:

  1. (1)

    Three prominent metrics are utilized to evaluate the performance of the algorithms (discussed in 5.1).

  2. (2)

    The minimum distance of the Pareto front set is utilized to choose the gold point from the ideal point in Euclidean two-dimensional distances (discussed in 5.2).

  3. (3)

    The compromised distance of the Pareto front set is utilized to analyze the minimum distance from the ideal point in Euclidean n-dimensional distances (discussed in 5.4).

  4. (4)

    The proposed algorithm has evaluated the improvement of real train operations. The main target is to evaluate the average passenger waiting time and the total number of train operations on one cycle in line Tehran–Mashhad of Iran (discussed in 6).

All algorithms' execution parameters are mentioned in Table 3. Noteworthy, these algorithms with different conditions and parameters are compared with each other.

Table 3 List of algorithm parameters

The Taguchi method is one of the best statistical methodologies, which uses finding the minimum number of experiments to tune the parameters of the algorithms [46]. It is utilized to find the optimum values of the algorithm's parameters mentioned in Table 3, which is implemented by Minitab software V16.2.4.4. Two approaches are provided to analyze the experimental results and select the most appropriate parameters: 1. standard method, which calculates the effect of factors, and variance analysis and 2. Signal-to-noise-ratio (SNR), which calculates dispersion relative to a specific value [47]. In this article, the standard method is utilized to analyze the experiments.

5.1 Algorithm evaluation by metrics

In the research communities, there is no comprehensive agreement about the metrics for evaluating multi-objective optimization algorithms [48]. Researchers are more inclined to metrics such as Diversity, Spacing, Mean Ideal Distance, and Pareto front [49]. Convergence speed and exploration–exploitation balance are critical to a role in local optimum points search [50]. To maintain exploration efficiency in hybrid algorithms, the widely utilized number of neighborhoods is five [38].

Table 4 shows the statistical results of several Pareto fronts, metrics of diversity, spacing, and MID. The NSGA-II-SA, MOEA-D, MOEA_HSS, and MOGWO hybrid algorithms improve the diversity and convergence more than the NSGA-II algorithm.

Table 4 the algorithm evaluation metrics

Diversity Diversity shows the feasible solutions in the Pareto front, which is diversity is an essential factor for the solution search in the Pareto front. Diversity allows the genetic algorithms to explore multiple regions of the solution space and avoid getting stuck in the local optimum. The genetic algorithm may converge too quickly to suboptimal solutions with a population that has low diversity [51]. The NSGA-II has the low performance of other algorithms by a value of 0.2879. The MOEA-D, MOEA-HSS, and MOGWA have values of 0.4987, 0.4361, and 0.4797 in the same range, respectively. The NSGA-II-SA has a high diversity with a value of 0.5341. The NSGA-II-SA has more diversity to generate elite offspring and better convergence than other algorithms.

$$ {\mathbf{Diversity}}_{{{\text{NSGA}} - {\text{II}}}} < {\mathbf{Diversity}}_{{{\text{MOEA}} - {\text{HSS}}}} < {\mathbf{Diversity}}_{{{\text{MOGWA}}}} < {\mathbf{Diversity}}_{{{\text{MOEA}} - {\text{D}}}} < {\mathbf{Diversity}}_{{{\text{NSGA}} - {\text{II}} - {\text{SA}}}} $$

Spacing The spacing metric is an excellent indicator to evaluate the approximation uniformity and the distribution quality of solutions in the Pareto optimal set. The distance indicator measures the degree of deviation toward getting stuck in local optima. An algorithm with a lower spacing metric has a better performance than others. [49].According to the results of the three algorithms, NSGA-II, MODEA-D, and MOGWA have values of 0.6658, 0.6234, and 0.6432 in a similar range, respectively. They have indicated almost the same performance. The NSGA-II-SA has the highest spacing metric of 0.5927 the MOEA-HSS has the least amount spacing metric of 0.7872, which have a difference acted more than 24%. Their results represent significant differences in spacing matric.

The distance indicator measures the degree of deviation toward getting stuck in local optima.

$$ {\mathbf{Spacing}}_{{{\text{MOEA}} - {\text{HSS}} }} > {\mathbf{Spacing}}_{{{\text{NSGA}} - {\text{II}} }} > {\mathbf{Spacing}}_{{{\text{MOGWA}}}} > {\mathbf{Spacing}}_{{{\text{MOEA}} - {\text{D}} }} > {\mathbf{Spacing}}_{{{\text{NSGA}} - {\text{II}} - {\text{SA}}}} $$

MID The Mean Ideal Distance is the size of how far each Pareto front individual is from the ideal point in the objective space. MID is calculated as the average distance between each solution on the Pareto front and the ideal solution, which allows for a quick assessment of how well-suited the solutions are toward satisfying all objectives simultaneously. The main goal in multi-objective optimization problems is to achieve the lowest MID possible so that Pareto front solutions successfully equilibrate all objectives as closely as possible [52]. The results of the MID show that the NSGA-II-SA has the highest MID metric of 2.8321, and the NSGA-II-SA has the least amount spacing metric of 5.7623, which have a difference acted more than 24%. The MOEA-D, MOEA-HSS, and MOGWA have values of 3.2476, 3.1181, and 3.3265 in the analogous range, respectively, which has the same act.

$$ {\mathbf{MID}}_{{{\text{NSGA}} - {\text{II}}}} > {\mathbf{MID}}_{{{\text{MOEA}} - {\text{D}}}} > {\mathbf{MID}}_{{{\text{MOGWA}}}} > {\mathbf{MID}}_{{{\text{MOEA}} - {\text{HSS}} }} > {\mathbf{MID}}_{{{\text{NSGA}} - {\text{II}} - {\text{SA}}}} $$

5.2 Pareto front data normalization

The proposed model is a train scheduling problem with two objective functions and discrete variables. The numerical amounts of the objective functions (Z1, Z2) differ considerably in numerical value. The result of objective functions (Z1, Z2) has normalized to have a good understanding, better display, analysis, and create a balance between the amounts [44]. In formulas (15)–(18), the variables \({z}_{k}^{max}\) and \({z}_{k}^{min}\) are maximum and minimum for each objective function, respectively, n is the number of the Pareto front population, and k is the number of the objective function [53].

$$ \theta = \left\{ {\theta_{1} ,\theta_{2} , \ldots ,\theta_{n} } \right\} \to {\text{the pareto front set}} \,n \ge 2 $$
(15)
$$ Z_{k} \left( \theta \right) = \left\{ {z_{k} \left( {\theta_{i} } \right)} \right\} \quad k = 1,2 \,{\text{and}} \,i = 1,2, \ldots , n $$
(16)
$$ z_{k}^{\max } = \max Z_{k} \left( { \theta } \right) \,{\text{and}}\, z_{k}^{\min } = \min Z_{k} \left( \theta \right) \quad k = 1,2 $$
(17)
$$ \overline{z}_{k} \left( {\theta_{i } } \right) = \frac{{z_{k} \left( {\theta_{i } } \right) - z_{k}^{min} }}{{z_{k}^{max} - z_{k}^{min} }} 0 \le \overline{z}_{k} \left( {\theta_{i } } \right) \le 1 \; k = 1,2 \;i = 1,2, \ldots ,n $$
(18)

5.3 Gold point on Pareto front set

Multiple Attribute Decision Making (MADM) and the Gold Point method are various approaches to decision-making. The MADM approach is utilized for decisions with several inconsistent attributes. The attributes include recognizing, analyzing, and evaluating conflicting alternatives with different criteria for the decision-maker [54]. The MADM approach includes allocating weights criteria and detecting the best alternative based on each criterion alternative [55]. The gold point approach is for a specific MADM method that selects one optimal solution from several alternatives. In summary, the MADM is an approach to decision-making with several attributes and criteria. The gold point is a specific technique of the MADM that chooses the optimal decision among several alternatives.

The results of traditional multi-objective optimization problems include several local optimal points [9]. The railway operators need to select one program of the local optimum points. Therefore, the single-objective optimization problems method can select one local optimal point between the results of a multi-objective optimization problem. The gold programming of the decomposition methods is an approach to select a locally optimal point [56]. In the first step, the point (0, 0) is chosen as the ideal point with the lowest value for the objective functions. In the next step, the value of Euclidean norm (2- norm) for each local optimum point relative to the ideal point is calculated by formula (19). The third step is to select an individual with the lowest Euclidean norm with formula (20).

Finally, the gold point GP is chosen as an optimal solution among optimal solutions [12], and the train schedule is extracted for five algorithms.

$$ Euclidean \, norm \, \left( {2 - \, norm} \right): \overline{\user2{Z}}_{ } \left( { \theta_{ } } \right)_{2} = \sqrt {\mathop \sum \limits_{i = 1}^{n} \overline{z}_{ } \left( { \theta_{i } } \right)^{2} } $$
(19)
$$ GP = \min \overline{\user2{Z}}\left( \theta \right)_{2} $$
(20)

In Fig. 9, the classical two-dimensional representation scatters negative correlation plot is drawn with normalized Pareto front data, the gold point, and the Euclidean distance (Ed) for five algorithms. A scatter plot helps visually understand the relationship between objective functions. The middle points in the Pareto front are usually closer to the ideal solution, and the population of the Pareto front near the gold point has the highest diversity and the lowest crowding distance [51, 57, 58].

Fig. 9
figure 9

The scatter negative correlation plot of five algorithms

5.4 Algorithms analysis

Multi-Criteria Decision Making (MCDM) is used for problems with multiple criteria [59], which (TOPSIS) and analytic hierarchy process (AHP) are two different methods of MCDM. In the AHP method, a pairwise comparison matrix has been utilized to calculate the weights of each criterion and alternative, which is determined based on subjective factors, relative priorities, and experts' opinions [60]. The AHP method has some weaknesses including relative value of options, large computational resource requirement, more time, mental nature, more pairwise comparisons, and application of tastes by experts [61, 62]. As another MCDM approach, the TOPSIS model is to rank the solutions according to relative closeness to the ideal points. Therefore, the best alternative should have less distance from the ideal positive point and most distance from the ideal negative point [63]. The primary difference between TOPSIS and AHP is that TOPSIS is based on finding the best alternative among all solutions, whereas AHP selects the best based on the importance of criteria.

It is a fact that passenger satisfaction will be improved by reducing the waiting time and increasing the number of trains, which result in increasing the train operating costs. On the other hand, passenger dissatisfaction will be increased by reducing the number of trains, which results in decreasing train operating costs. These types of challenges are permanent challenges in daily railway operations. This subsection analyzes the compromise solution by the TOPSIS method. The compromise results are shown in Table 5.

Table 5 the compromise solution by algorithms

Figures 10, 11 and 12 show a big picture of the compromise between problem objective functions with different weights of TOPSIS. One normal matrix or weighted scale is created to assign weight to each criterion. The criteria weights are determined by expertise and range from zero to one, with a total sum of one. In this study, the weights are chosen for the objective functions with equal values (0.3, 0.7), (0.5, 0.5), and (0.7, 0.3).

Fig. 10
figure 10

The compromised solutions of algorithms for weight = (0.7, 0.3)

Fig. 11
figure 11

The compromised solutions of algorithms for weight = (0.5, 0.5)

Fig. 12
figure 12

The compromised solutions of algorithms for weight = (0.3, 0.7)

In Fig. 10, the MOEA-D and MOGWA algorithms have a more orientation to the negative ideal point than others. In addition, the NSGA-II, NSGA-II-SA, and MOEA-HSS algorithms have a more inclination to the positive ideal point. The algorithm MOGWA has a more interquartile range IQR and diagonal to a negative ideal solution than other algorithms. The MOGWA and MOEA_D have a more tendency to the negative ideal point, respectively.

In Fig. 11, all algorithms have solutions with a relatively symmetrical distribution, but the MOGWO tends to a negative ideal solution and negative skewness. The best compromise has happened with minimum IQR by NSGA-II-SA.

In Fig. 12, the algorithm NSGA-II has a more considerable interquartile range IQR, diagonal to a negative ideal solution, and negative skewness than other algorithms. The algorithm MOEA-HSS has widespread IQR, balance IQR, and a little negative skewness. The algorithm NSGA-II-SA has widespread and balanced IQR and more positive skewness. MOEA-D and MOGWA algorithms have almost similar performance.

The algorithms may choose different alternatives to the optimal solution by different parameters. The NSGA-II-SA has obtained the best-compromised solution than other algorithms with values (0.5, 0.5).

6 Numerical experiment

In this section, through numerical calculation and evaluation, the experimental results of the proposed model will be presented. The model is implemented and executed utilizing the developed NSGA-II-SA algorithm and four algorithms, namely NSGA-II, MOEA-HSS, MOEA/D, and MOGWO with the simulation environment specification presented in Table 6.

Table 6 Simulation environment specification

6.1 Case study

The case study has been analyzed by comparing the actual train schedule of the Iranian railway and an optimal solution with the NSGA-II-SA algorithm with the lowest Ed (mentioned in 5.3) to evaluate the proposed model by a real dataset. It should be noted that real datasets are the same in all algorithms. Figure 13 shows the double-track with a high-traffic railway route of the Tehran–Mashhad, Iran. This part of the double-track railway network is 924 km long, has 50 stations, and has 58 trains’ movements in 24 h.

Fig. 13
figure 13

Part of the map of the Iran railway [64]

Figure 14, shows the scheduling graph of high-traffic passenger trains on the Tehran–Mashhad route for 24 h.

Fig. 14
figure 14

Real scheduling graph Tehran–Mashhad route

Train traffic management needs to be more efficient in dispatch, avoiding the cause of congestion and operation complexity. The train schedule is run from 00:30 AM to 5:00 PM on the outbound route and 9:30 AM to 12:00 PM on the return. The train dispatch density causes passenger presence crowding in stations in the limited time of 24 h. The turnaround interval is reduced between two successive trains due to crowding and traffic of turning trains at the first and last stations. As a result, the time interval must be decreased between two sequence trains. This approach causes operation density and increases traffic management risk.

Figure 15 shows the train schedule graph by the NSGA-II-SA algorithm on the same day and with the same data. Trains are scheduled to run from 00:00 AM to 10:30 PM on the outbound route and from 00:30 AM to 11:00 PM on the return route on one cycle. According to constrain (5), the difference between the maximum and minimum time traveling between two stations i, i + 1 is proportional to the dispatch of trains during a 24-h cycle. Constrain (6) and (7) control the difference between the maximum and minimum safe distance between two consecutive trains \({{h}_{u}}^{min}, {{h}_{u}}^{max}\). The distribution of passengers \({a}_{iu}\left(p\right)\) spread in the station, service management, and the time required for train rotation \({{t}_{u}}^{T}\) are expanded on the first and last station. This algorithm has improved traffic management risk, congestion, and operation density of trains in dispatch.

Fig. 15
figure 15

Modeling graph NSGA-II-SA Tehran–Mashhad route

6.2 Scenarios analysis of the proposed deterministic-based approach

The genetic algorithm (GA) is one of the population-based stochastic algorithms, with a randomly initialized population, random crossover operator, and random mutation operator. The SA is a stochastic algorithm with a global search that tries to escape the trap of local optima points [26]. First, consider the uncertainty of passenger demands on the model. Suppose the uncertain passenger demands occur with known probability under seasons. Passenger arrivals follow a Poisson distribution, where the arrival intensity λu (14) of each passenger arriving is considered the same at variant stations, at different hours of the day, in each season [75]. In this case study, the probability of the number of passengers entering the station in scenarios U (spring, summer, fall, and winter) is considered to be 0.65, 0.85, 0.55, and 0.45, respectively.

Table 7 illustrates the normalized values of the first and second objective functions, Average Passenger Waiting time (APW), and the Total Number of Trains (TNT) in four scenarios. The highest and lowest amount of function Z1 occurred in the algorithm the NSGA-II and NSGA-II-SA with APW = 15.3 and APW = 10.09 under scenario summer and winter, respectively. The highest and lowest amount of function Z2 is in the algorithm the MODE-HSS and NSGA-II-SA with TNT = 57 and TNT = 49 under scenario spring and winter, respectively. The best optima solution is in the algorithm NSGA-II-SA with APW = 10.09 and TNT = 54 under scenario summer. The worst case is in the algorithm MOEA-HSS with APW = 14.52 and TNT = 57 under scenario spring. The difference between the most optimal solution and the worst case in terms of the average passenger waiting time and the total number of trains is 43.9% and 5.55%, respectively.

Table 7 the algorithms result based on four scenarios (spring, summer, fall, winter)

Figure 16 shows the average travel time of the NSGA-II-SA algorithm on gold point (GP) and the actual train scheduling between stations for all trains on one route. The average passenger waiting time has been calculated for the NSGA-II-SA algorithm and the real train scheduling 10.09 and 10.90 min, respectively. This figure shows that the average passenger waiting time decreases, and passengers will experience an average waiting time of 7.35% decrease based on function (1).

Fig. 16
figure 16

Average travel time between stations on one route

Table 8 includes the result of the algorithms and the actual train schedule of the Tehran–Mashhad trains.

Table 8 Comparison of algorithms and real scheduling one cycle

This table is regulated based on a GP and Ed of the Pareto front and includes the average passenger waiting time and the total number of trains in an operating cycle. The percentage improvement has been compared to five algorithms with the real train schedule. Finally, the NSGA-II-SA algorithm has improved the average passenger travel time and the number of train operations on one cycle by 7.34% and 6.89%, respectively.

7 Conclusions and future research

Train schedule is a stochastic multi-objective optimization problem with many constrains and parameters, by which achieving the ideal solution is not practical in a short time due to passenger satisfaction. Passenger satisfaction will be improved by reducing the waiting time and increasing the number of trains, which result in increasing the train operating costs. In contrast, passenger dissatisfaction will be increased by reducing the number of trains, decreasing train operating costs. These types of challenges are permanent challenges in daily railway operations. To alleviate these challenges, a stochastic multi-objective optimization model for train schedules in Tehran–Mashhad, Iran, route has been developed. NSGA-II meta-heuristic algorithm was developed, and as a hybrid model, its combination with the simulated annealing (SA) algorithm has been presented. Then, the proposed model has been solved simultaneously with five algorithms, namely NSGA-II, MOEA-HSS, MOEA-D, MOGWA, and NSGA-II-SA multi-objective meta-heuristic algorithms, and compared based on several metrics. After, the TOPSIS method has been utilized to analyze the compromised solutions on the Pareto front set. In the case study, the gold programming of the decomposition method was utilized to select only one optimal solution among all Pareto front optimal solutions. Following, the improvement percentage has been calculated for the actual operation. The main goal is to study NSGA-II-SA algorithm efficiency as the optimal solution superior for reducing the average passenger waiting time and the total number of train cycles in one operational period.

For future research, it is suggested to combine multi-objective meta-heuristic algorithms with artificial intelligence, machine learning techniques, or game theory. The following changes can be applied in the objective functions according to the train operation requirements. The parameters include passenger satisfaction, the probability of accidents, delays caused by a locomotive or wagon deficiency, train speed and weight, fuel consumption reduction, rescheduling, and moving blocks. In the case study, the results of the multi-objective optimal problem can be solved by several single-objective optimization methods until the best compromise solution is gained with the real operations railway during the different 24 h.