1 Introduction

For more than 40 years exam timetabling has been a particularly well studied domain across the Artificial Intelligence and Operational Research communities. This is due to its importance in many academic institutions worldwide. However, much of the research has been aimed at developing methodologies that produce the best quality timetables for a single problem (Qu et al. 2009b). A more recent direction in this field, represented by some work on hyper-heuristics, aims to raise the level of generality of search methodologies to create algorithms that operate effectively over a range of problems. A hyper-heuristic can be seen as a heuristic to choose heuristics (Burke et al. 2003). In this case, the low-level heuristics represent the search space. In timetabling, the low-level heuristics can be categorised as heuristics which construct a timetable or heuristics which perform certain moves to improve a constructed timetable. This paper presents a random iterative hyper-heuristic approach which uses improvement low-level heuristics. This approach was tested on the Toronto benchmark and the second International Timetabling Competition (ITC2007) exam timetabling problems. It was found to generalise well over the two datasets. Furthermore, very competitive results have been produced against other approaches in the literature.

The following section presents a brief description of the benchmarks. Sections 2.3 and 2.4 provide an overview on different approaches developed in the exam timetabling domain. Furthermore, Sect. 2.5 concentrates on hyper-heuristic approaches. A random iterative hyper-heuristic to improve timetables is proposed in Sect. 3. An adaptive methodology to select low-level heuristics and the results obtained are presented in Sect. 4. The future extensions of this work are summarised in Sect. 5.

2 Exam timetabling

2.1 The Toronto benchmark

An exam timetabling problem consists of the allocation of a set of exams to a given set of timeslots. The generated timetable must satisfy the hard constraints of the problem which are the requirements that cannot be violated. For example, no one student must be scheduled to sit two exams during the same period. A timetable which meets all the hard constraints given is called a feasible timetable. A timetabling problem can also have a set of soft constraints. Soft constraints are constraints that can be violated and the level of violation determines the quality of the timetable generated. An example of a soft constraint is that there must be a certain number of periods between two exams taken by the same student. Therefore, high quality timetables contain the least number of soft constraint violations. The Toronto benchmark problem is well known in the exam timetabling community since it was first introduced by Carter et al. (1996) in 1996. Over the years, a slightly different version was made available and used to test some approaches in the literature. The characteristics of the two versions of this dataset are presented in Table 1. The problem has one hard constraint. This says that conflicting exams (with common students) must not be assigned to the same time slot. In addition, there is one soft constraint. This says that conflicting exams should be spread throughout the timetable as much as possible. The objective is to minimise the sum of proximity costs given as:

where

  • w i =24−i is the cost of assigning two exams with i time slots apart. Only conflicting exams with common students and which are four or less time slots apart are considered as violations

  • n is the number of students involved in the conflict

  • S is the total number of students in the problem

This problem has been used to test and compare many approaches in the literature. Recently, a more constrained set of benchmarks was made available as part of the International Timetabling Competition (ITC2007) (McCollum et al. 2010). The next section describes the ITC2007 dataset in detail.

Table 1 Characteristics of the two versions of the Toronto Benchmark datasets

2.2 The international timetabling competition (ITC2007) dataset

The ITC2007 exam timetabling track could be considered to be a complex and a more practical dataset in comparison to the Toronto benchmark. This is due to the larger number of constraints it contains. A full description of the problem and the evaluation function can be found in McCollum et al. (2010). In addition, the characteristics which define the instances are summarised in Table 2. The problem consists of the following:

  • A set of timeslots covering a specified length of time. The number of timeslots and their durations are provided.

  • A set of exams which should be allocated to the timeslots.

  • A list of students enrolled in each exam.

  • A set of rooms with different capacities.

  • A set of additional hard constraints.

  • A set of soft constraints and their associated penalties.

Table 2 Characteristics of the ITC2007 dataset

In comparison to the Toronto benchmark, the ITC2007 dataset has more than one hard constraint. The hard constraints can be outlined as follows:

  • No student sits more than one exam at the same time.

  • The capacity for each individual room should not be exceeded at a given period.

  • Period lengths should not be violated.

  • All period related hard constraints need to be satisfied e.g. exam A after exam B.

  • All room related hard constraints need to be satisfied e.g. exam A must use room X.

The soft constraint violations can be summarised as follows:

  • Two Exams in a Row The number of occurrences where a student sits two exams in a row on the same day.

  • Two Exams in a Day The number of occurrences where a student sits two exams on the same day. If the exams are back to back then this is considered as a Two Exams in a Row violation to avoid duplication.

  • Period Spread The exams have to be spread a certain number of timeslots apart.

  • Mixed Durations The number of occurrences where exams of different durations are assigned to the same room.

  • Larger Exams Constraint The number of occurrences where the largest exams are scheduled near the end of the exam session. The number of the largest exams and the distance from the end of the exam session are specified in the problem description.

  • Room Penalty The number of times where certain rooms, which have an associated penalty, are used.

  • Period Penalty The number of times where certain timeslots, which have an associated penalty, are used.

2.3 Exam timetabling approaches for the ITC2007 dataset

A three phased approach was developed by Muller (2008) to solve the problems in the ITC2007 exam timetabling track. The first phase consists of an Iterative Forward Search algorithm to find a feasible solution. Hill climbing is used to find the local optima in the second phase. Finally, a Great Deluge Algorithm is applied to further explore the search space.

Gogos et al. (2008) proposed a method which used a GRASP (Greedy Randomised Adaptive Search Procedure) approach. In the construction phase, five orderings of exams based on various criteria are generated. Tournament selection is used to select exams until they are all scheduled. A backtracking strategy using a tabu list is employed as required. In the improvement phase, Simulated Annealing is used. Finally, room allocations are arranged using integer programming in the third phase.

Atsuta et al. (2008) used a constraint satisfaction solver which employed tabu search and iterated local search. The solver differentiates between the constraints and their corresponding weights during the computation to improve performance. De Smet (2008) also incorporated local search techniques in a solver called Drools, an Open-Source Business Rule Management System (http://www.jboss.org/drools/).

Pillay (2008) introduced a biologically inspired approach which is analogous to cell behaviour. The exams are initially ordered using the saturation degree heuristic and scheduled sequentially in the available “cells” i.e. timeslots. If more than one timeslot is available, the slot which causes the least overall constraint violation is chosen. Rooms are chosen using the best fit heuristic. If a conflict occurs before all the exams are scheduled, the timetable is rearranged to reduce the soft constraint violation. This is described as cell division. If the overall soft constraint violation is not improved without breaking the hard constraints, cell interaction occurs. The timeslots are swapped in this process to remove hard constraint violations. The process continues until a feasible solution is achieved. Finally, the contents of cells having equal durations are swapped to improve the solution. This is called cell migration.

McCollum et al. (2009) proposed a two phased approach where an adaptive heuristic is used to achieve feasibility during the first phase. The second phase improves the solution through the employment of a variant of the Great Deluge Algorithm.

Finally, Pillay (2008) presented an evolutionary algorithm based hyper-heuristic using three different chromosome representations. An initial population is created and iteratively improved by applying the processes of evaluation, selection and recreation.

2.4 Exam timetabling approaches for the Toronto benchmark

Abdullah et al. (2009) presented a hybridisation of an electromagnetic-like mechanism (EM) and the Great Deluge algorithm. The technique estimates the quality of the solution and calculates a decay rate at every iteration of the search. These values depend on a force value calculation using the EM approach. This approach produced the best quality solution for four of the instances. An approach which uses a sequential construction method, employed by Caramia et al. (2008), to assign exams in the least number of timeslots was able to produce the best quality timetables for another four of the Toronto benchmark instances. It uses a greedy scheduler to obtain a feasible solution. A penalty decreaser and trader are then applied to improve the quality of the constructed solution. Burke et al. (2010) introduced an approach which combines a variable neighbourhood search with a genetic algorithm which produced the best quality solution for one of the Toronto instances. In addition, Burke and Bykov (2008) proposed a method where a hill-climber compares the candidate solution with a solution produced a couple of iterations back instead of the current solution. This was called the “late acceptance criteria” and it produced the best quality solutions for another two instances. The results obtained by these approaches are presented in Sect. 4.1.

Recently, some new methods were investigated to automatically find the best heuristic to solve a set of instances. This has led to the introduction of hyper-heuristics. The next section summarises some of the hyper-heuristic methods applied to the exam timetabling domain.

2.5 Hyper-heuristics in exam timetabling

A hyper-heuristic can be seen as a method to choose low-level heuristics depending on the problems in hand. Furthermore, it could be used to adapt or tune heuristics and meta-heuristics. Hyper-heuristics in exam timetabling can be categorised, according to the low-level heuristics they use, into two types as follows:

  1. 1.

    Hyper-heuristics with constructive low-level heuristics

  2. 2.

    Hyper-heuristics with improvement low-level heuristics

A Tabu search approach was developed by Burke et al. (2007) to optimise a search space of heuristic sequences comprised of two or more low-level heuristics. This work was extended in later research by Qu et al. (2009a) to construct heuristic sequences which produce feasible timetables. The combinations are then analysed to find distribution patterns of low-level heuristics. This leads to the generation of sequences which are adaptively adjusted to construct better timetables. In addition, hybridisations of the graph based hyper-heuristic with local search methods were investigated in Qu and Burke (2009). The formal definition of the search problem of hyper-heuristics, where low-level heuristics represent the search space, has also been provided in Qu and Burke (2009).

Asmuni et al. (2004) used fuzzy logic to combine two out of three graph colouring heuristics. The idea was to combine the two heuristics into a single value which calculates the difficulty of allocating an exam to a timeslot. The exams are ordered using this value and are scheduled in order. Furthermore, the approach was extended to tune the fuzzy rules instead of keeping them fixed (Asmuni et al. 2009).

Ersoy et al. (2007) developed an approach called the hyperhill-climber where a hyper-heuristic is embedded in a memetic algorithm. The aim of this hyper-heuristic was to select the best hill-climber to apply or decide the best order in which hill-climbers are executed. In addition, Pillay and Banzhaf (2007) created another approach where genetic programming was used to evolve hyper-heuristics.

Biligan et al. (2007) presented different heuristic selection methods and acceptance criteria for hyper-heuristics in exam timetabling. Finally, a different method of combining heuristics was presented by Pillay and Banzhaf (2009). The low-level heuristics were combined hierarchically and applied simultaneously instead of being applied sequentially.

3 A hyper-heuristic with low-level improvement heuristics

Several low-level heuristics can be used to improve a timetable with varying quality. The different low-level heuristics used could be considered as different methods for escaping from local optima. However, the order in which exams are moved and the type of moves performed play an important role in finding the best quality solution. An initial feasible solution is constructed using the Largest Degree heuristic where the exams in the ordering are assigned to the timeslot causing the least penalty. In the case where there is more than one timeslot with the lowest penalty, one of them is randomly chosen. Our objective is to analyse the performance of the different low-level heuristics used to minimise the penalty incurred from a constructed solution. In addition, we test the effect of using different orderings for the exams causing penalties in the solution. Finally we develop an adaptive approach which orders the exams causing violations and automatically selects the best heuristic to use for each exam to produce an improvement.

3.1 The low-level heuristics

In this paper we investigate the effect of using different low-level heuristics or neighbourhood operators to improve timetables. A combination of two improvement low-level heuristics is used in our approach. The following is a list of the heuristics investigated:

  1. 1.

    Move Exam (ME): This heuristic selects an exam and reassigns it to the timeslot causing the least penalty.

  2. 2.

    Swap Exam (SE): This heuristic selects an exam and tries to swap it with a scheduled exam leading to the least penalty timetable.

  3. 3.

    Kempe Chain Move (KCM): This is similar to the SE heuristic but is more complex as it involves swapping a subset of conflicting exams in two distinct timeslots. This neighbourhood operator was successfull in some earlier studies (Burke et al. 2010; Thompson and Dowsland 1996).

  4. 4.

    Swap Timeslot (ST): This heuristic selects an exam and swaps all the exams in the same timeslot with another set of exams in a different timeslot. After testing all the timeslots, the swap producing the least penalty timetable is applied.

3.2 The random iterative hyper-heuristic

The study presented in this paper takes a similar approach to that presented in Qu et al. (2009a) where a random iterative hyper-heuristic generates heuristic sequences of different quality to solve the benchmark problem mentioned in Sect. 2.1. Instead of being used to construct solutions, the heuristic sequences are used here to improve constructed feasible solutions by rescheduling exams causing penalties. Algorithm 1 presents the pseudo-code of this random iterative hyper-heuristic. The process starts by constructing an initial feasible solution. Since the initial solution constructed affects the improvement process, a random largest degree graph colouring heuristic which orders exams according to the number of conflicts each exam has with others is used (Burke et al. 2010). This allows us to compare our approach to other approaches in the literature which use a similar method in construction. At every iteration, the exams causing violations in the constructed solution are identified and a random sequence of moves is generated. A move is the application of one of the low-level heuristics described in Sect. 3.1. The sequence of moves is then applied to the sequence of exams as they are unscheduled one by one. Only moves that improve the current solution are accepted. If a move does not improve the solution, it is skipped and the exam stays in its current position. A sequence is discarded if an improvement is not obtained after the whole sequence is applied.

Algorithm 1
figure 1

The pseudo-code of the random iterative hyper-heuristic with low-level improvement heuristics

This approach was applied to four instances (hec92 I, sta83 I, yor83 I and tre92) of the Toronto benchmark exam timetabling problems described in Sect. 2.1 for off-line learning of the best heuristic hybridisations and the order of execution leading to the best improvement. These instances were chosen as they vary in size and cover a range of conflict densities (see Table 1). After running this process for (e×50) times, where e is the number of exams causing soft constraint violations in the constructed solution, a set of sequences and the penalties of their corresponding solutions are obtained for further investigation on the effectiveness of the different heuristics used. Note that, only 50 samples were collected for each rate of hybridisation as we found that using more samples does not improve the quality of the final outcome at this point. Finally, an adaptive approach was developed and applied to the Toronto benchmark. Furthermore, to test the generality of the approach, it was applied to the ITC2007 exam timetabling track. The approach is presented in Sect. 4.

3.3 Analysis of hybridising improvement low-level heuristics

In order to clearly observe the effect of the different low-level heuristics in improving solutions, the heuristic sequences generated consist of two heuristics. We use the Kempe chain move heuristic as the basic heuristic in the sequences as it has proved to be successful in previous work (Burke et al. 2010; Thompson and Dowsland 1996). The Kempe chain move involves swapping a subset of exams in two distinct timeslots making sure that a hard constraint violation does not occur. The rest of the heuristics (ME, SE and ST) are randomly hybridised into the list of KCM.

The random sequences are generated with different percentages of hybridisation by inserting n ME, SE or ST, n=[1,…,e] in the sequences. For each hybridisation of KCM with either ME, SE or ST, 50 samples are obtained for each amount of hybridisation. Duplicate sequences are discarded and another sequence is generated instead. The sequences are re-initialised in each iteration to lead the search to explore a wider area of the search space at this stage.

We applied this approach to four instances of the Toronto benchmark exam timetabling problems (Carter et al. 1996). Table 3 presents the results obtained using ME, SE and ST in a hybridisation with KCM as well as a comparison against using KCM only.

Table 3 Results using KCM without a hybridisation and with several different moves

It is observed that using a Kempe chain only produces the worst results. After introducing other heuristics in a hybridisation with the Kempe chain moves, better results were obtained. Another observation from Table 3 is that swapping timeslots and performing Kempe chain moves produces the best improvement for all the problems. One possible reason may be that swapping timeslots allows the search to be more diverse and to sample different areas of the search space to find good solutions more quickly. In addition, no obvious trends could be obtained on the amount of ST hybridisation within the best heuristic sequences. However, it is observed in all the sequences leading to the best timetables, the ST heuristic is randomly distributed within the sequence and the percentage of hybridisation is less than 50 %.

3.4 Variations of orderings of the exams causing a penalty

To analyse the effect of ordering the unscheduled exams causing a soft constraint violation in a previous solution, we decided to test different orderings while using the Kempe Chain and swapping timeslot hybridisation stated in the previous section. After the exams causing violations are identified, they are ordered first before being reassigned to a timeslot. Several orderings can be used to guide the search as follows:

  • Largest Degree (LD): The exams are ordered decreasingly according to the number of conflicts each exam has with others.

  • Largest Weighted Degree (LWD): The exams are ordered similarly to LD but weighted according to the number of students involved in the conflict.

  • Saturation Degree (SD): The exams are ordered increasingly according to the number of remaining timeslots available to assign them without causing conflicts. In the case where ties occur, LWD is used as a tie breaker. From our previous work it was shown that SD produces the best results when LWD is used to break ties in the ordering (Burke et al. 2011).

  • Largest Penalty (LP): The exams are ordered decreasingly according to the penalty they incur in the current solution.

  • Random Ordering (RO): The exams are ordered randomly.

Table 4 presents the results of applying different orderings to the unscheduled exams, then running a random heuristic sequence of KCM and ST to assign them in better timeslots.

Table 4 Results of hybridising KCM with ST using different orderings of the exams causing a soft constraint violation. The notation X tb Y means heuristic Y is used to break ties in heuristic X

As shown in Table 4, we found that using SD and breaking any ties in the ordering using LWD produced the best results. This is because SD orders the unscheduled exams according to the number of timeslots available to assign them without causing conflicts. Therefore, the chances of moving exams at the top of the SD list and finding better timeslots for them become higher. Ordering the exams according to the penalty they incur was seen to be the second best ordering followed by LWD. The effect of using LD and RO was the same as randomly choosing an exam to reschedule.

A t-test is also carried out to give an indication if the results using SD, LP and LWD are significantly different. Tables 5, 6 and 7 summarise the p-values of the t-tests carried out between the results of different orderings, which are significantly different in all the cases.

Table 5 t-test on the results from ordering exams causing violations using SD and LP
Table 6 t-test on the results from ordering exams causing violations using SD and LWD
Table 7 t-test on the results from ordering exams causing violations using LP and LWD

4 Adaptive selection of low-level heuristics for improving exam timetables

Algorithm 2 presents the initialisation stage of the adaptive approach. The exams causing a penalty are first identified and are unscheduled. They are then put in a list and ordered using SD. Random heuristic sequences are generated using KCM and ST to reschedule the exams. The sequences are then applied to the ordered exams and the corresponding solutions are saved. Note that, only 10 sequences are generated for each rate of hybridisation to be able to adhere to the time limitation and compare our results with the best in the literature for the ITC2007 dataset. Furthermore, limiting the number of sequences generated in this stage makes it easier to analyse and observe any trends in the sequences generating the best results.

Algorithm 2
figure 2

The pseudo-code of the initialisation stage of the adaptive hyper-heuristic with low-level improvement heuristics

The above observations indicate that the best solutions were obtained when ordering the exams causing violations using SD, and rescheduling them using either a Kempe-chain move or swapping timeslots. It was also observed that the heuristic sequences producing the top 5 % results used the same move for the majority of the exams (i.e. the same heuristic appears at the same position in more than 75 % of the sequences). Therefore, we developed an intelligent approach that performs an analysis of the best 5 % of the sequences produced to generate a new set of sequences. The new set of sequences obtained better results for all the problem instances. The adaptive approach was tested and showed to be effective and comparable with the best approaches in the literature.

Algorithm 3 presents the pseudo-code of the approach which hybridises ST with KCM in two stages. The process is presented as follows:

  1. 1.

    In the first stage, the best 5 % of heuristic sequences are collected and analysed. If the same heuristic is used at the same position for more than 75 % of the heuristic sequences, then it is stored. Otherwise the position is kept empty. Note that, we also tested collecting more than 5 % of the best sequences to analyse them. However, the behaviour was random since no trends were seen. Therefore, to guarantee the effectiveness of the approach, the heuristic was chosen for a certain position if it appears in 75 % of the best 5 % heuristic sequences collected.

  2. 2.

    In the second stage, the empty positions are randomly assigned as KCM or ST. (n×5) sequences for the large problems (uta92 I, uta92 II, car91 and car92) and (n×10) sequences for the small problems are generated, respectively. The generated sequences are then applied to the problem.

Algorithm 3
figure 3

Adaptive generation of heuristic sequences hybridising KCM and ST

4.1 The Toronto benchmark results

We tested this approach on the Toronto benchmark exam timetabling problems and present the results in Tables 8 and 9. The average computational time for each stage across the instances is also presented for 30 runs on a Pentium IV machine with a 1 GB memory. In addition, the number of exams causing a penalty is presented in the tables.

Table 8 Results from the adaptive improvement Hyper-heuristic (AIH) approach on the Toronto Benchmark dataset.
Table 9 Contd. Results from the adaptive improvement Hyper-heuristic (AIH) approach on the Toronto Benchmark dataset

The best results stated in the literature are presented in Table 10. These include the hybridisation of an electromagnetic-like mechanism (EM) and the Great Deluge algorithm employed by Abdullah et al. (2009), the hill-climbing with a late acceptance strategy implemented by Burke and Bykov (2008), the variable neighbourhood search incorporating the use of genetic algorithms used by Burke et al. (2010) and the sequential construction method developed by Caramia et al. (2008). These algorithms are briefly discussed in Sect. 2.4.

Table 10 Best results obtained by the Adaptive Improvement Hyper-heuristic (Algorithm 3) compared to the best approaches in the literature on the Toronto Benchmark

The results obtained indicate the generality of our approach to different constructed timetables regardless of the size. We also make a comparison with other hyper-heuristics which produced the best results in the literature in Table 11. In comparison with the graph-based hyper-heuristic in Burke et al. (2007), our approach performs better in all the cases reported. In addition, it performs better in 8 out of 11 cases in comparison with the hyper-heuristics investigated in Pillay and Banzhaf (2009) and Qu and Burke (2009). Finally, it performs better in 10 out of 11 cases compared to the tabu search hyper-heuristic investigated in Kendall and Mohd Hussin (2005). Only the problems presented in Table 11 were compared to other results since the results for the other instances in Table 1 were not reported in the literature.

Table 11 Best results obtained by the Adaptive Improvement Hyper-heuristic (AIH) compared to other hyper-heuristics approaches in the literature on the Toronto Benchmark

4.2 The international timetabling competition (ITC2007) results

To test the generality of our approach, we applied it to the ITC2007 exam timetabling dataset. The initial solution is constructed by ordering the exams according to their saturation degree. The exams are assigned a random timeslot in the situation where more than one timeslot is available. After a feasible solution is constructed the Adaptive Improvement Hyper-heuristic in Algorithm 3 was applied to the constructed solution. To allow a fair comparison with the reported competition results, the approach was run for the same amount of time using 11 distinct seeds for each instance. Table 12 presents the results we obtained in comparison with the best in the literature. The description of the approaches used for comparison is presented in Sect. 2.3. We do emphasise that the objective here is not to beat the best reported results but to demonstrate the generality of our approach to different problems with different constraints. A dash in the table means that no feasible solution was obtained.

Table 12 Best results obtained by the Adaptive Improvement Hyper-heuristic (AIH) compared to the best approaches in the literature on the ITC2007 dataset

The Extended Great Deluge in McCollum et al. (2009) obtained the best results for 5 out of the 8 instances. However, the approach was run for a longer time as it was developed after the competition. In the competition, the best results for all the 8 instances were reported in Muller (2008) using a three phased approach. The GRASP used in Gogos et al. (2008) produced the second best results.

In comparison with other hyper-heuristic techniques, our approach was able to produce better results in only one instance when compared to the evolutionary algorithm based hyper-heuristic presented in Pillay (2010). However, it was stated that these experiments did not adhere to the time limitation imposed by the competition.

In comparison to the Constraint Based Solver developed in Atsuta et al. (2008), our approach performed better in 3 out of the 8 instances. The approach using the Drools solver in De Smet (2008) obtained feasibility for only 5 instances. Our approach outperformed it as we were able to gain feasibility for all the 8 instances. This demonstrates the generality of our approach to solving exam timetabling problems. Finally, our approach performed better on 6 of the 8 instances in comparison with the biologically inspired approach proposed in Pillay (2008).

5 Conclusions

The study presented in this paper implements a hyper-heuristic approach which adaptively adjusts heuristic combinations to achieve the best improvement on constructed timetables. An investigation is made of the low-level heuristics used and the order in which exams causing soft constraint violations are rescheduled. The analysis is performed on a set of four benchmark instances of differing difficulty in an off-line learning process. It is shown that, of the heuristics tried, the timeslot swapping heuristic is the best to combine with Kempe chains. In addition, better solutions are produced when ordering the exams causing a soft constraint violation using Saturation Degree and breaking any ties with Largest Weighted Degree. Based on the output of the learning process, an adaptive approach which analyses and adjusts some randomly generated sequences is implemented and applied to the rest of the instances. Furthermore, the hyper-heuristic approach is applied to a more constrained dataset, and showed to produce very competitive results compared to other approaches in the literature on both datasets.

Future research directions include performing improvements during the timetable construction stage instead of performing the improvements at the end of the construction. Using hybridisations of more than two low-level heuristics could also be investigated. Finally, the approach investigated in this paper can be applied to course timetabling.