Abstract: As the disadvantages of the conventional ant colony optimization, such as convergence speed and local optimal value, an improved multi-agent ant colony optimization for reactive power optimization method is proposed. The improved sort-weighted ant colony optimization is combined with multi-agent system to improve convergence speed and calculation accuracy and to avoid falling into local optimal value effectively. Improved algorithm is applied to reactive power optimization in IEEE 30-node systems, the simulation results verify the validity of the algorithm.
INTRODUCTION
Reactive power optimization which searches the optimal solution in the given constraint situation, plays an important role in power quality, network loss and voltage stability. Scholars have introduced the intelligent optimization algorithm to the field of reactive power optimization to overcome the shortcomings of classic algorithms in recent decades. Evolutionary Programming (EP) which opened up the process of intelligent optimization algorithm was first proposed by L.J. Fogel ect. Evolutionary Strategies (ES) and Genetic Algorithm (GA) presented in 1970s promoted the further development of intelligent algorithm. In 1995, Differential Evolution Algorithm (DE) was discussed by R. Storm and K. Price and at the same time, another branch of intelligent algorithm-swarm intelligent algorithm, especially Ant Colony Optimization (ACO) and Particle Swarm Optimization (PSO) (Zhang, 2011) are applied widely. Although different optimization algorithms are put forward, they have different scope of application. ACO has no systemic mathematical foundation and analysis methods and the disadvantages such as complex calculation, local optimal value, slowly convergence. Much improvement has been done for ACO. For example, ACO (Tan, 2012) based on chaos theory utilizes the ergodicity of chaotic search and avoids to fall into local optimal value, ACO combined with Immune Algorithm (Bao and Yang, 2007) to improve single search mechanism which enhances global search ability at the expense of the operation time.
A method integrates Multi-Agent System (MAS) with improved ACO to overcome local optimal value and increase convergence speed is discussed in the study which can achieve fast and accurate optimization.
MODEL OF REACTIVE POWER OPTIMIZATION
The mathematical model of reactive power optimization consists of objective function and constraint conditions. The constraint conditions include equality constraints and inequality constraints.
Several constraints such as network loss, voltage quality, reactive compensation capacity and economic benefits ect., should be considered to establish reactive power optimization objective function. Considering these factors comprehensively, the objective function is as follows:
(1) |
where, F is distribution electric energy loss, n is the total number of branch; Gij is conductance on the i-j branch; δi δj, respectively stand for phase angles of nodei and node j; uiuj stand for voltages of nodei and nodej.
As the power flow calculation is used to not only test the results of optimization but also is the main component of calculation program in some optimization algorithm, so the power flow equations are treated as equality constraint conditions, it can be expressed as follows:
(2) |
where, Pi and Qi are the injected active power and reactive power of nodei separately; Ui and Uj stand for voltages of node i and node j; Gij and Bij are real part and imaginary part of element in the ith row and the jth column in system node admittance matrix.
The reactive power of generator, compensation capacity of reactive power compensation devices and variable ratio of on-load tap changer are selected as controlling variables. Voltages of load node are selected as state variables.
The inequality constraints of control variable are:
(3) |
The inequality constraints of state variable are:
(4) |
where, QGi is reactive power of nodei generator; QGimax and QGimin are upper and lower limit of QGi; QCi is the compensation capacity of the ith reactive power compensation device; upper and lower limit of QCi are QCimax and QCimin; Ti is variable ratio of the ith on-load tap changer; Timax and Timin are upper and lower limit of Ti; uimax and uimin are upper and lower limit of ui, voltage of nodei.
MULTI-AGENT Ant COLONY OPTIMIZATION
Conventional ant colony optimization: Ant Colony Optimization (ACO) is a new simulated evolutionary algorithm which simulates real ants foraging behavior in the nature. But the difference with real ant system is ACO has memory function (Shi, 2008) by exchanging message and cooperation among ants which can get the optimal value of optimized problem which can be equivalent to the shortest path of real ant foraging from the nest to food source.
Traveling salesman problem (TSP) is taken as an example to explain the basic principle of ACO. Initial parameters m is the number of ants, n is the scale of the city, the distance between cityi and cityj is dij (i,j = 1,2,...,n), the concentration of pheromone that is between the path of cityi and cityj at t moment is τij(t). Ant k (k = 1,2,...,m) decides to visit next city according to the concentration of the pheromone between two cities. PKij(t) is probability of ant k from cityi to cityj at t moment, the expression is:
(5) |
where,ηij (t) is enlightening function, its value is 1/dij and stands for the expectation from cityi to cityj, allowk expresses set of cities which will be visited by ant k and reduces from n-1 to 0 as time goes on. Pheromone enlightening factor is α and self-enlightening factor is β. The greater the value that is defined between α and β, the greater transfer function between two cities is.
The pheromone among the path of different cities, also gradually disappears when the ants release pheromone. P is defined as volatile coefficient (0<P<1), then real-time concentration of pheromone among the path of cities can be updated when all ants complete the cycle:
(6) |
(7) |
(8) |
where, the increment of pheromone concentration on the searching path (i,j) of ant k isΔτKij; the sum of pheromone concentration on the searching route (i,j) of all ants is Δτij; Q is constant and it is amount of pheromone which is released by ant after circulation; Lkis the length of route that is passed by ant k.
The basic process of ACO is shown in Fig. 1.
Improved ant colony optimization: Some scholars introduced the concept of sorting of GA into ACO (Ren, 2006) which improved the algorithm convergence speed in TSP.
Fig. 1: | Basic process of ACO |
Based on the thought, the weight coefficient of ACO is adjusted and applied to reactive power optimization to solve convergence speed.
The main idea of improved algorithm is as follows. The paths which are obtained at the end of one ant optimal circulation are sorted according to the length. An ants contribution of updating pheromone depends on the length of the path. The shorter circulation generates the path, the greater contribution is made by ants to pheromone. The rules of pheromone updating are modified by adding a weighted coefficient λK(1-N/Nmax). All the ants have corresponding contribution for global pheromone updating, where the better ants are placed a high value and the effect of worse ants is also weakened during the process of optimization.
In improved algorithm, the path and the global pheromone are updated according to the equation:
(9) |
where, λis a constant, 0<λ<1; Q is also a constant expressing the amount of pheromone which is released by ant after circulation; k is the No. of optimal ants; Lk is the length of route that is passed by ant k; N and Nmax are respectively current iteration No. and the maximum iteration No.
When the initial parameters is same, the results of common ACO and improved ACO are shown in Fig. 2.
The advantages of improved ACO over conventional ACO can be summarized from Fig. 2 as follows:
• | The improved ACO is able to achieve global optimal value faster than conventional ACO, while satisfying all the power system operation constraints and convergence conditions |
• | Searching curve of improved ACO is more smooth |
• | The improved ACO avoids trapping in local optimal value |
Fig. 2: | Comparing results of common ACO and improved ACO |
Fig. 3: | Grid structure of agent |
Multi-agent Ant colony (MAS) optimization: Multiple-Agent system can solve complex optimal problem by collaborating of each agent. The Agent modules having different functions complete coordinately the task through communication and cooperation with greater flexibility and adaptability (Dagdeviren et al., 2011).
Multi-Agent Ant Colony Optimization (MACO) has multiple characters which combine MAS with improved ACO. Any Agent is equivalent to an ant in ACO and it also has a best adaptive value which is decided by optimization.
Agent environment: The survival environment of agent can be simplified to a grid structure in constrained condition. Agent must carry pheromone in ACO and has corresponding response.gridstructure is demonstrated in Fig. 3, S stands for the size of grid structure and S*S is ant colony scale.
Fig. 4: | Local environment of non-boundary agent |
Fig. 5: | Local environment of boundary agent |
Local environment and competitive-cooperative operations: Every agent competes and cooperates with its neighbors, then takes corresponding action strategies independently combined with ant colony searching. Thus, the local environment of Agent is significant (Zhao and Cao, 2005). For the non-boundary Agenti,j, its neighbors are shown in
Figure 4 and can be expressed as:
(10) |
where, i and j are both positive integer and I, j0 (1,2,3 ,S)
For boundary Agent, namely i,j = 1 or S. Its neighbors are shown in Fig. 5. The value of row index:
(11) |
In the same way, the value of column index:
(12) |
Agents are enlightened by pheromone which is left by ants and cooperate and compete with their neighbors. The agent with optimal adaptive value keeps its original location, then exchanges pheromone to all ants at the same time. This algorithm increases the speed of transfer and can search the global optimal solution quickly.
Self-iearning operation: Self learning operation search in small scope and dont introduce ant colony searching mechanism. Self-learning operation is only used to optimal agent to enhance searching speed effectively.
Optimizing steps of MACO: The steps of the improved algorithm is as follows:
• | Step 1: Set the upper and lower limit of initial parameters and constraints and define the specified control parameters and the maximum number of iterations in the algorithm |
• | Step 2: Create the Agent survival environment,gridstructure and initial Agent, the initial number of iteration is 0 |
• | Step 3: Evaluate the adaptive value of each Agent by Newton-Raphson power flow algorithm |
• | Step 4: Every agent and its neighbors are competitive and cooperative, then updating Agents in the whole environment |
• | Step 5: Carry out the improved ACO in the MAS and update position of each Agent in the solution space for the second time afterwards |
• | Step 6: Evaluate the adaptive value of Agent again; |
• | Step 7: Search the optimal adaptive value of Agent, (search the minimum network loss value in this article) and update the location of Agent in the solution space for the third time according to self-learning operation of Agent |
• | Step 8: times of iteration add 1 |
• | Step 9: Judge whether the termination conditions is reached, that is, maximum number of iterations or convergence condition; if not, jump to Step 3. Otherwise, terminate the iterations and output the optimal value of optimization |
EXAMPLE ANALYSIS
The improved algorithm is applied to IEEE 30-node system (Liu, 2010) analysis to verify the feasibility and validity. IEEE 30-node system is shown in Fig. 6.
The power reference in the system is SB = 100MVA. There are 6 generators, 38 branches, 4 transformers, 9 sets of parallel capacitor and 21 load nodes. The total load active power is Ptotal = 2.834 and reactive power is Qtotal =1.262. Parameters are set in Table 1 and 2.
In the initial conditions, 3PG and 3QG are 2.8691 and 1.3807 respectively by power flow calculation. Network loss is Ploss which equal to 0.035. The initial parameters of the improved algorithm are set as follows. Total environment size is S = 10, namely, ant population scale is 100, the largest number of iterations is Nmax= 50, pheromone enlightening factor is α = 1, self-enlightening factor is β = 2, volatile coefficient is P = 0.1, weighting coefficient is λ = 0.3, self-learning environment size is s = 4, searching radius of self-learning sR = 0.4.
Fig. 6: | IEEE 30-node system |
Table 1: | Generator parameters |
Table 2: | Transformer ratio parameters |
Table 3: | Contrastive results of optimization |
The comparison result of MACO and GA in the same initial conditions is demonstrated in Table 3.
In Table 3, the power network is optimized by MACO, network loss is 0.017 which is much less than GA and initial. Compared with GA, there is a 17% increase in network loss descent rate and the computing time is shorter. By means of power flow calculation, the reactive power of balance bus become much less than before. The results fully show effectiveness of MACO by taking economic condition and the result of optimization into consideration.
CONCLUSION
The reactive power optimization is a important technique to ensure power system running safely and economically. An improved multi-agent ant colony algorithm is discussed to realize reactive power optimization which overcomes the disadvantage of slow convergence and falling into local optimal solution by modifying the pheromone weight. The improved algorithm is combined with multi-agent system and applied to IEEE-30 node system, the results demonstrate that the computation speed and accuracy are all increased.
ACKNOWLEDGMENTS
The work is sponsored by Science and Technology Research Projects in Education Department of Heilongjiang Province (No. 12531062) and Chang Jiang Scholar Candidates Program for Provincial Universities in Heilongjiang (No. CJHB2012005).