HOME JOURNALS CONTACT

Information Technology Journal

Year: 2014 | Volume: 13 | Issue: 8 | Page No.: 1561-1566
DOI: 10.3923/itj.2014.1561.1566
Reactive Power Optimization in Power System Based on Improved Multi-agent Ant Colony Optimization
Jianhong Yao, Lingyu Zhang, Daxing Sun and Jianbai Song

Abstract: As the disadvantages of the conventional ant colony optimization, such as convergence speed and local optimal value, an improved multi-agent ant colony optimization for reactive power optimization method is proposed. The improved sort-weighted ant colony optimization is combined with multi-agent system to improve convergence speed and calculation accuracy and to avoid falling into local optimal value effectively. Improved algorithm is applied to reactive power optimization in IEEE 30-node systems, the simulation results verify the validity of the algorithm.

Fulltext PDF Fulltext HTML

How to cite this article
Jianhong Yao, Lingyu Zhang, Daxing Sun and Jianbai Song, 2014. Reactive Power Optimization in Power System Based on Improved Multi-agent Ant Colony Optimization. Information Technology Journal, 13: 1561-1566.

Keywords: reactive power optimization, intelligent algorithm and Power system

INTRODUCTION

Reactive power optimization which searches the optimal solution in the given constraint situation, plays an important role in power quality, network loss and voltage stability. Scholars have introduced the intelligent optimization algorithm to the field of reactive power optimization to overcome the shortcomings of classic algorithms in recent decades. Evolutionary Programming (EP) which opened up the process of intelligent optimization algorithm was first proposed by L.J. Fogel ect. Evolutionary Strategies (ES) and Genetic Algorithm (GA) presented in 1970s promoted the further development of intelligent algorithm. In 1995, Differential Evolution Algorithm (DE) was discussed by R. Storm and K. Price and at the same time, another branch of intelligent algorithm-swarm intelligent algorithm, especially Ant Colony Optimization (ACO) and Particle Swarm Optimization (PSO) (Zhang, 2011) are applied widely. Although different optimization algorithms are put forward, they have different scope of application. ACO has no systemic mathematical foundation and analysis methods and the disadvantages such as complex calculation, local optimal value, slowly convergence. Much improvement has been done for ACO. For example, ACO (Tan, 2012) based on chaos theory utilizes the ergodicity of chaotic search and avoids to fall into local optimal value, ACO combined with Immune Algorithm (Bao and Yang, 2007) to improve single search mechanism which enhances global search ability at the expense of the operation time.

A method integrates Multi-Agent System (MAS) with improved ACO to overcome local optimal value and increase convergence speed is discussed in the study which can achieve fast and accurate optimization.

MODEL OF REACTIVE POWER OPTIMIZATION

The mathematical model of reactive power optimization consists of objective function and constraint conditions. The constraint conditions include equality constraints and inequality constraints.

Several constraints such as network loss, voltage quality, reactive compensation capacity and economic benefits ect., should be considered to establish reactive power optimization objective function. Considering these factors comprehensively, the objective function is as follows:

(1)

where, F is distribution electric energy loss, n is the total number of branch; Gij is conductance on the i-j branch; δi δj, respectively stand for phase angles of nodei and node j; uiuj stand for voltages of nodei and nodej.

As the power flow calculation is used to not only test the results of optimization but also is the main component of calculation program in some optimization algorithm, so the power flow equations are treated as equality constraint conditions, it can be expressed as follows:

(2)

where, Pi and Qi are the injected active power and reactive power of nodei separately; Ui and Uj stand for voltages of node i and node j; Gij and Bij are real part and imaginary part of element in the ith row and the jth column in system node admittance matrix.

The reactive power of generator, compensation capacity of reactive power compensation devices and variable ratio of on-load tap changer are selected as controlling variables. Voltages of load node are selected as state variables.

The inequality constraints of control variable are:

(3)

The inequality constraints of state variable are:

(4)

where, QGi is reactive power of nodei generator; QGimax and QGimin are upper and lower limit of QGi; QCi is the compensation capacity of the ith reactive power compensation device; upper and lower limit of QCi are QCimax and QCimin; Ti is variable ratio of the ith on-load tap changer; Timax and Timin are upper and lower limit of Ti; uimax and uimin are upper and lower limit of ui, voltage of nodei.

MULTI-AGENT Ant COLONY OPTIMIZATION

Conventional ant colony optimization: Ant Colony Optimization (ACO) is a new simulated evolutionary algorithm which simulates real ants foraging behavior in the nature. But the difference with real ant system is ACO has memory function (Shi, 2008) by exchanging message and cooperation among ants which can get the optimal value of optimized problem which can be equivalent to the shortest path of real ant foraging from the nest to food source.

Traveling salesman problem (TSP) is taken as an example to explain the basic principle of ACO. Initial parameters m is the number of ants, n is the scale of the city, the distance between cityi and cityj is dij (i,j = 1,2,...,n), the concentration of pheromone that is between the path of cityi and cityj at t moment is τij(t). Ant k (k = 1,2,...,m) decides to visit next city according to the concentration of the pheromone between two cities. PKij(t) is probability of ant k from cityi to cityj at t moment, the expression is:

(5)

where,ηij (t) is enlightening function, its value is 1/dij and stands for the expectation from cityi to cityj, allowk expresses set of cities which will be visited by ant k and reduces from n-1 to 0 as time goes on. Pheromone enlightening factor is α and self-enlightening factor is β. The greater the value that is defined between α and β, the greater transfer function between two cities is.

The pheromone among the path of different cities, also gradually disappears when the ants release pheromone. P is defined as volatile coefficient (0<P<1), then real-time concentration of pheromone among the path of cities can be updated when all ants complete the cycle:

(6)

(7)

(8)

where, the increment of pheromone concentration on the searching path (i,j) of ant k isΔτKij; the sum of pheromone concentration on the searching route (i,j) of all ants is Δτij; Q is constant and it is amount of pheromone which is released by ant after circulation; Lkis the length of route that is passed by ant k.

The basic process of ACO is shown in Fig. 1.

Improved ant colony optimization: Some scholars introduced the concept of sorting of GA into ACO (Ren, 2006) which improved the algorithm convergence speed in TSP.

Fig. 1: Basic process of ACO

Based on the thought, the weight coefficient of ACO is adjusted and applied to reactive power optimization to solve convergence speed.

The main idea of improved algorithm is as follows. The paths which are obtained at the end of one ant optimal circulation are sorted according to the length. An ant’s contribution of updating pheromone depends on the length of the path. The shorter circulation generates the path, the greater contribution is made by ants to pheromone. The rules of pheromone updating are modified by adding a weighted coefficient λK(1-N/Nmax). All the ants have corresponding contribution for global pheromone updating, where the better ants are placed a high value and the effect of worse ants is also weakened during the process of optimization.

In improved algorithm, the path and the global pheromone are updated according to the equation:

(9)

where, λis a constant, 0<λ<1; Q is also a constant expressing the amount of pheromone which is released by ant after circulation; k is the No. of optimal ants; Lk is the length of route that is passed by ant k; N and Nmax are respectively current iteration No. and the maximum iteration No.

When the initial parameters is same, the results of common ACO and improved ACO are shown in Fig. 2.

The advantages of improved ACO over conventional ACO can be summarized from Fig. 2 as follows:

The improved ACO is able to achieve global optimal value faster than conventional ACO, while satisfying all the power system operation constraints and convergence conditions
Searching curve of improved ACO is more smooth
The improved ACO avoids trapping in local optimal value

Fig. 2: Comparing results of common ACO and improved ACO

Fig. 3: Grid structure of agent

Multi-agent Ant colony (MAS) optimization: Multiple-Agent system can solve complex optimal problem by collaborating of each agent. The Agent modules having different functions complete coordinately the task through communication and cooperation with greater flexibility and adaptability (Dagdeviren et al., 2011).

Multi-Agent Ant Colony Optimization (MACO) has multiple characters which combine MAS with improved ACO. Any Agent is equivalent to an ant in ACO and it also has a best adaptive value which is decided by optimization.

Agent environment: The survival environment of agent can be simplified to a grid structure in constrained condition. Agent must carry pheromone in ACO and has corresponding response.“grid”structure is demonstrated in Fig. 3, S stands for the size of grid structure and S*S is ant colony scale.

Fig. 4: Local environment of non-boundary agent

Fig. 5: Local environment of boundary agent

Local environment and competitive-cooperative operations: Every agent competes and cooperates with its neighbors, then takes corresponding action strategies independently combined with ant colony searching. Thus, the local environment of Agent is significant (Zhao and Cao, 2005). For the non-boundary Agenti,j, its neighbors are shown in

Figure 4 and can be expressed as:

(10)

where, i and j are both positive integer and I, j0 (1,2,3…,S)

For boundary Agent, namely i,j = 1 or S. Its neighbors are shown in Fig. 5. The value of row index:

(11)

In the same way, the value of column index:

(12)

Agents are enlightened by pheromone which is left by ants and cooperate and compete with their neighbors. The agent with optimal adaptive value keeps its original location, then exchanges pheromone to all ants at the same time. This algorithm increases the speed of transfer and can search the global optimal solution quickly.

Self-iearning operation: Self learning operation search in small scope and don’t introduce ant colony searching mechanism. Self-learning operation is only used to optimal agent to enhance searching speed effectively.

Optimizing steps of MACO: The steps of the improved algorithm is as follows:

Step 1: Set the upper and lower limit of initial parameters and constraints and define the specified control parameters and the maximum number of iterations in the algorithm
Step 2: Create the Agent “survival” environment,“grid”structure and initial Agent, the initial number of iteration is 0
Step 3: Evaluate the adaptive value of each Agent by Newton-Raphson power flow algorithm
Step 4: Every agent and its neighbors are competitive and cooperative, then updating Agents in the whole environment
Step 5: Carry out the improved ACO in the MAS and update position of each Agent in the solution space for the second time afterwards
Step 6: Evaluate the adaptive value of Agent again;
Step 7: Search the optimal adaptive value of Agent, (search the minimum network loss value in this article) and update the location of Agent in the solution space for the third time according to self-learning operation of Agent
Step 8: times of iteration add 1
Step 9: Judge whether the termination conditions is reached, that is, maximum number of iterations or convergence condition; if not, jump to Step 3. Otherwise, terminate the iterations and output the optimal value of optimization

EXAMPLE ANALYSIS

The improved algorithm is applied to IEEE 30-node system (Liu, 2010) analysis to verify the feasibility and validity. IEEE 30-node system is shown in Fig. 6.

The power reference in the system is SB = 100MVA. There are 6 generators, 38 branches, 4 transformers, 9 sets of parallel capacitor and 21 load nodes. The total load active power is Ptotal = 2.834 and reactive power is Qtotal =1.262. Parameters are set in Table 1 and 2.

In the initial conditions, 3PG and 3QG are 2.8691 and 1.3807 respectively by power flow calculation. Network loss is Ploss which equal to 0.035. The initial parameters of the improved algorithm are set as follows. Total environment size is S = 10, namely, ant population scale is 100, the largest number of iterations is Nmax= 50, pheromone enlightening factor is α = 1, self-enlightening factor is β = 2, volatile coefficient is P = 0.1, weighting coefficient is λ = 0.3, self-learning environment size is s = 4, searching radius of self-learning sR = 0.4.

Fig. 6: IEEE 30-node system

Table 1: Generator parameters

Table 2: Transformer ratio parameters

Table 3: Contrastive results of optimization

The comparison result of MACO and GA in the same initial conditions is demonstrated in Table 3.

In Table 3, the power network is optimized by MACO, network loss is 0.017 which is much less than GA and initial. Compared with GA, there is a 17% increase in network loss descent rate and the computing time is shorter. By means of power flow calculation, the reactive power of balance bus become much less than before. The results fully show effectiveness of MACO by taking economic condition and the result of optimization into consideration.

CONCLUSION

The reactive power optimization is a important technique to ensure power system running safely and economically. An improved multi-agent ant colony algorithm is discussed to realize reactive power optimization which overcomes the disadvantage of slow convergence and falling into local optimal solution by modifying the pheromone weight. The improved algorithm is combined with multi-agent system and applied to IEEE-30 node system, the results demonstrate that the computation speed and accuracy are all increased.

ACKNOWLEDGMENTS

The work is sponsored by Science and Technology Research Projects in Education Department of Heilongjiang Province (No. 12531062) and Chang Jiang Scholar Candidates Program for Provincial Universities in Heilongjiang (No. CJHB2012005).

REFERENCES

  • Zhang, X., 2011. Study on Intelligent Optimization Algorithm and its Applications for Reactive Power Optimization in Power System. Southwest Jiaotong University, Sichuan, China


  • Tan, S.B., 2012. Research on Reactive Power Optimization in Distribution Networks Based on Chaos Ant Colony Algorithm. South China University of Technology, Guangzhou, China


  • Bao, Y.D. and W. Yang, 2007. Application of new intelligent algorithm to reactive power optimization. Mod. Electr. Power, 24: 16-19.


  • Shi, F., 2008. Thirty Cases Analysis about Intelligent Algorithm Based on Matlab. Beihang Press, Beijing, China, pp: 205-207


  • Ren, R.C., 2006. A Rank-Weight-Based Version of Ant Colony Optimization. Dalian Maritime University, Dalian, China


  • Dagdeviren, O., I. Korkmaz, F. Tekbacak and K. Erciyes, 2011. A survey of agent technologies for wireless sensor networks. IETE Tech. Rev., 2: 168-184.
    Direct Link    


  • Zhao, B. and Y. Cao, 2005. A multi-agent particle swarm optimization algorithm for reactive power optimization. Comput. Sci. Electri. Eng., 25: 1-7.


  • Liu, M.B., 2010. Optimal Power Flow Calculation in Large Power System. Science Press, Beijing, China, pp: 381-384

  • © Science Alert. All Rights Reserved