Abstract
This paper is concerned with the short-term load forecasting (STLF) in power system operations. It provides load prediction for generation scheduling and unit commitment decisions; therefore precise load forecasting plays an important role in reducing the generation cost and the spinning reserve capacity. In order to improve the precision of electric power system load forecasting, the hybrid algorithm which combines improved genetic algorithm with radial basis function (RBF) neural network is used in short-term load forecasting of electric power system in this paper. In the model, disruptive selection strategy, adaptive crossover and mutation probability were adopted to improve population diversity during iterative process and prevent premature convergence. The improved genetic algorithm and gradient descent method were mixed for interactive computing, and the hybrid algorithm was used for RBF learning. The model was applied to the actual system. The results were compared with traditional RBP algorithm and offered a high forecasting precision.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
STLF is a very important part of power system operation, and it is also an integral part of the energy management system (EMS). STLF is the basis of optimal operation for power system. Prediction accuracy has a significant impact on safety, quality, and economic performance of the power system. Therefore, it is very important to search more suitable STLF method in order to maximize the prediction accuracy.
On the one hand, the power system load relates to many complex factors, and it is non-linear, uncertain and random. On the other hand, RBF neural network has the abilities of strong self-learning and complex non-linear function fitting, so it is suitable for load forecasting problems. But studies show that there are still many problems to be solved on the RBF neural network algorithm, particularly the parameters identify of RBF network, such as the number of hidden layer nodes, the center and width value in the RBF activation function for each node, and connection weights between the hidden layer nodes and output layer node. These parameters have great impact on the RBF network learning speed and performance. If improper parameters are selected, network convergence will slow, or even result in the network does not converge. In this paper, an improved genetic algorithm is used to optimize the RBF network, and the optimized network is used to forecast power load. Case analysis and calculation show that the method has high accuracy and good applicability.
2 Radial Basis Function Neural Network
2.1 RBF Network Structure
Radial basis function network is a local approximation network, generally including three layers (n inputs, m hidden node, p output). The structure is shown in Fig. 1.
Basis function of RBF network is commonly used Gaussian function, which can be expressed as
Where: \( {\phi_i}(x) \) is the output of the i-th hidden layer node; X is the input sample, \( \mathrm{ x} = ( {{\mathrm{ x}}_1},{{\mathrm{ x}}_2},\ldots,\ {{\mathrm{ x}}_{\mathrm{ n}}} ){{}^{\mathrm{ T}}} \); \( {{\mathrm{ c}}_{\mathrm{ i}}} \) is the center of the Gaussian kernel function of the i-th hidden layer node, having the same number of dimensions with X; \( {\sigma_{\mathrm{ i}}} \) is a variable for the i-th hidden layer node, called normalization constant or the base width [1, 2]. RBF network output is a linear combination of the output of hidden layer nodes.
2.2 The Learning Algorithm of Network Parameters
In the section, Gradient descent method is used to study the center \( {{\mathrm{ c}}_{\mathrm{ i}}} \) and the width parameters \( {\sigma_{\mathrm{ i}}} \) of RBF network. For the convenience of discussion, consider the case that the output layer has only one node. Substitute Eq. 1 into Eq. 2:
Let the network desired output is \( {y^d}(x) \), the network energy function can be expressed as:
Let \( f({x^j}) \) substitute into Eq. 4, we can get:
Let the sample size is L, then
Remember
Then Eq. 6 becomes
\( {w_i} \) is consider as a constant when you learn center values and width parameters, and you can get the central value and width parameter updating formula which can be represented by:
In Eqs. 9 and 10, λ, β are the learning efficiency of the central value and width parameters. If the formula (8) is substituted into the formula (9) and formula (10), we can get:
According to the input samples, weights of the output layer can be calculated by using the least squares algorithm of system identification theory. In this paper, the learning algorithm for connection weights between the hidden layer and output layer can be expressed as:
In the formula: \( y_{{_k}}^d \) is the output to expect; l is the number of iterations; η is the learning rate, general 0 <η <2, to ensure iterative convergence.
3 RBF Network Training Based on Improved Genetic Algorithm
The core of RBF network design is to determine the number of hidden nodes and the central values, and other parameters of basis functions. We will design a neural network to meet the target error as small as possible to ensure that the generalization ability of neural networks. Genetic algorithm (GA) is a randomized search method which simulates biological evolution. This paper presents an improved genetic algorithm (IGA) [3]; the starting point for the algorithm is described as follows. First: to maintain population diversity and prevent premature; second: to improve local search ability of GA; Third: speed up the search; Fourth: to reduce the chance of getting into local extreme value.
3.1 Encoding and Initial Population Generation
That is to say, the RBF network hidden nodes number m, each hidden node center parameters \( {{\mathrm{ c}}_{\mathrm{ i}}} \) and the width parameters \( {\sigma_{\mathrm{ i}}} \) are compiled chromosome, and the collection of these parameters for the network are treated as an individual. In the initialization phase, initial population is generated by completely random method.
3.2 Select Options
Based on the deviation degree of this population, this article defines a selection operator which can bring a diversity of species, which is described as follows:
For a given fitness measure f, so
Among them, \( \bar{f}(X)=\frac{1}{N}\sum\limits_{k=1}^N {f({X_k})} \) is the population average fitness; N is the population size, \( f({X_j}) \) is the j-th individual’s fitness value in the group, \( U({X_j}) \) is the deviation degree between individual j and the group mean fitness. Disruptive selection chooses each individual according to the following probability formula:
From geometry, this means that the farther away from the average individual fitness, the higher the chance to be selected, thus corresponding with the fitness of the individual does not have a monotonic, can bring a greater diversity of species.
3.3 Cross Operating
The algorithm uses real number coding, therefore, the corresponding intersection operation can be realized by arithmetic crossover. Arithmetic crossover is defined as a linear combination of two individuals to generate a new breed of individual operations. We can set:
Among them, \( {X_1},{X_2} \) are two different individuals of the populations. We can take:
3.4 Mutation Operating
The adaptive mutation operator is used in this paper, and the specific description is as follows: First, if we randomly choose a component in the parent body vector \( x=({x_1},{x_2},\cdots, {x_n}) \), assumption it is the k-th, and then we randomly choose a number \( {{x^{\prime}}_k} \) instead of \( {x_k} \) in its definition interval \( [{a_k},{b_k}] \) to get mutated individuals \( y \), That is \( y=({x_1},{x_2},\cdots {{x^{\prime}}_k},\cdots, {x_n}) \), among them,
Where \( random(0,1) \) is the random number in the interval (0,1); \( \Delta (T,y)\in [0,y] \) is a random number obeying uniform distribution. As T decreases, the greater the likelihood \( \Delta (T,y) \) tends to 0. So that the algorithm searches large fitness individual in a small range and the small fitness individual in a large range, which makes the variation according to solution quality adaptively adjust the search area, which can obviously improve search capabilities [4].
The specific expression of the function \( \Delta (T,y) \) can be taken as:
Where \( r \) is random number of the interval [0, 1], \( \lambda \) plays a regulatory role of the local search area, and its value is generally 2–5. \( f(x) \) represents fitness of the individual \( x \), \( {f_{\max }} \) is the biggest fitness value of problem to be solved; Due to \( {f_{\max }} \) is difficult to determine in many problems, we can use rough upper or the largest fitness value of the current population.
3.5 Algorithm Realization
The RBF network structure optimization and parameter learning are carried out in two phases, namely training and evolution. First, randomly generate N individuals to form groups, We use a gradient descent to learn the center \( {{\mathrm{ c}}_{\mathrm{ i}}} \) and width parameters \( {\sigma_{\mathrm{ i}}} \) corresponding to each individual hidden nodes chromosomes in the network, and use the least squares to learn linear weight value \( {w_i} \) of the network; Secondly, We use genetic evolutionary algorithm to optimize hidden nodes, by alternating these two processes to obtain the minimum number of hidden nodes required to meet the error basis functions and have different width parameter of RBF network [5].
In order to use genetic algorithms to solve optimization problems for RBF network structure, Boolean vectors \( {U^T}=\left( {{u_{1, }}{u_{2, }}\ldots, {u_{M, }}} \right) \) \( {u_i}=\left\{ {0,1} \right\} \) is introduced. \( {u_i}=1 \) represents the corresponding hidden node exists; \( {u_i}=0 \) represents the corresponding hidden node does not exist. Each Boolean vectors \( {U^T} \) generate two chromosomes: one center parameters chromosomes, one width parameter chromosomes. Center parameters chromosome \( U_c^T \) and width parameters chromosome \( U_{\sigma}^T \) are coded by real number.
3.6 Example Application
In this paper, to predict the region May 12, 2005 24-point load, we uses 3 months the historical load data of the region Power Grid in 2005, The prediction results are shown in Table 1:
It can be seen from the table, the maximum relative error of the short-term load forecasting model based algorithm which proposed in this paper is 2.7, and the minimum relative error of it is 0.13. While the maximum relative error of prediction model based on RBF algorithm is 4.51, the minimum relative error is 0.87. It can be seen that the prediction model built in this article can better fit the mapping relationship between the loads; it has better prediction accuracy.
4 Conclusion
According to the deficiencies of RBF neural network and the premature shortcoming of genetic algorithm, this paper presented a radial basis function (RBF) neural network short term load forecasting model based on improved genetic algorithm. The model introduces the real-coded adaptive mechanism for the genetic algorithm. The selection strategy, adaptive crossover and mutation were improved, and its interaction with the gradient descent hybrid operation was used as the RBF network learning algorithm. The experimental results showed that the method can effectively improve the accuracy of load forecasting with good applicability.
References
Park J, Sandberg IW (1991) Universal approximation using radial basis-function networks[J]. Neural Comput 3(2):246–257
Roopesh R, Kumar Reddy, Ranjan Ganguli (2003) Structural damage detection in a helicopter rotor blade using radial basis function neural networks [J]. Smart Mater Struct 12(1):232–241
Stones J, Collinson A (2001) Power quality [J]. Power Eng J 15(2):58–64
Kandil MS, El Debeiky SM, Hasanien NE (2001) Overview and comparison of long term forecasting techniques for a fast developing [J]. Electr Power Syst Res 58(1):11–17
Zhang L, Bai YF (2005) Genetic algorithm trained radial basis function neural networks for modeling photovoltaic panels [J]. Eng Appl Artif Intel 18(7):833–844
Acknowledgements
A Project Supported by Scientific Research Fund of Hunan Provincial Science and Technology Programme, Project Number [2010FJ3157]
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Zhao, Y., Hu, H., Zhang, Y. (2014). The Application of Improved Genetic Algorithm Optimized by Radial Basis Function in Electric Power System. In: Wang, W. (eds) Mechatronics and Automatic Control Systems. Lecture Notes in Electrical Engineering, vol 237. Springer, Cham. https://doi.org/10.1007/978-3-319-01273-5_36
Download citation
DOI: https://doi.org/10.1007/978-3-319-01273-5_36
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-01272-8
Online ISBN: 978-3-319-01273-5
eBook Packages: EngineeringEngineering (R0)