IEEE Transactions on Power Apparatus and Systems, Vol. PAS-104, No. 11, November 1985
BAD DATA IDENTIFICATION METHODS IN POWER SYSTEM STATE ESTIMATION - A COMPARATIVE STUDY
M. Ribbens-Pavella
L. Mili (*)
la INTRODUCTION
a power system state e3timator softroutines, bad data identification is the last
but not least - satellite function. Its task is to
guarantee the reliability of the data base generated
through the estimator. Indeed, despitethe preprocessing
data validation techniques used to clear the data received at a control center, gross anomalies (suchas bad
data, modelling and parameter errors) may still exist
during estimation. To avoid corrupting the resulting
data base, it is of great importance that these anomalies are identified and further eliminated from the set
of measurements. This explains why the need for a function capable to identify bad data has been felt almost
simultaneously with the need for the state estimation
function itself. It also explains the number and diversity of research works carried out on the subject.
This paper aims at providing a comparative assessment of the "post-estimation" identification methods(1)
available today. More specifically it concentrates on
evaluating the techniques able to identify bad data
(BD), i.e. grossly erroneous measurements. These techniques are first classified, then explored and compared.
Three broad classes are distinguished : the class of
identification by elimrrnation (IBE) (1-14], that of the
non-quadratic criteria (NQC) [3,15-20], and the hypothesis testing identification (HTI) [211. The investigations are based upon both theoretical considerations
and practical experience. The latter has been acquired
through simulations performed on four different power
systems. The results reported here concern simulations
performed on the IEEE 30-bus system, with the three
possible types of multiple BD : noninteracting, interacting, and unidentifiable ones.
The paper is organized as follows. Section 2
gathers the material necessary for the intended exploration. The reader is supposed to be familiar at least
with state estimation and BD detection techniques; so
this Section focuses essentially on topological identifiability aspects and selection of identifiability
criteria. Section 3 gives a brief description of the
various identification methods within their corresponding categories,while Section 4 investigates further and
compares the three main methodologies. Finally, the exploration is completed and validated through the simulation results of Section 5.
In the list of
ware
2, MISCELLANIES
Somewhat hybrid, this Section groups the various
pieces of information necessary for the subsequent
developments. The degree of the authors' personal perception and interpretation goes increasing along the
paragraphs. Starting with definitions of the usual
symbols in 2.1, one is led up to some useful topological considerations and definitions in 2.3 and 2.4
and finally to the selection of relevant identifiability
criteria to be used in the comparative assessment of
the various identification methodologies.
2.1. STATE ESTIMATION: DEFINITIONS AN
SYMBOLS
model :
z=-i
4k;4'
through
the
(_% 1)
JL
ei
where
HA
ah/ax
N(0,WRWTT)=
the
-
weighted residual
rN
-
vector
with
rW
/R-rr
diag(WR)
(4)
(5)
J(e) =
rTR-U
=rWrW
(6)
non-
prob (ji|<X)
0018-9510/85/1100-3037$01.001985 IEEE
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3038
; is the statistical variable of concern (rwi,
J ) with meanvalue 1p and variance C2 ; X is the
detection threshold. Hence,detecting the presence of BD
requires that [241
|1| > X-N
where
rNi
or
(N7)
Definition.
Imax
in
[211);
02
a is the false
alarm probability
Wij
procedure :
(i) search for all measurements become critical after
elimination;
(ii) add these critical measurements to the list of the
measurements declared false;
(iii) determine the estimates which would be affectedby
possible errors on the critical measurements and join
this qualitative information to the final data base.
Step (i) can be carried out by simply comparingthe
lists of critical measurements before and after elimination.
The above procedure may apply to any identification
method which involves elimination of measurements.
therefore be used to determine the suspected measurements : these are measurements possessing normalized
residuals larger than the fixed threshold.
2.3. TOPOLOGICAL IDENTIFIABILITY OF BAD DATA
Given a set of BD it is interesting to determine
whether the measurement configuration is rich enough to
allow their proper identification.
Definition. A set of BD is said to be topoZogicaZZy
identifiable if their suppression does not cause
system's unobservability,
cient condition for proper identification; indeed numerical aspects have also to be taken into account.
A reliable identification procedure should be able
to recognize topologically unidentifiable BD; in such
cases, it should declare the problem unsolvable and warn
the operator against the lackof reliabilityof the available state estimate, rather than give unusable results.
LocacUzation oi the BV : ability to localize exactly the BD, or at least to furnish a list of suspected
measurements which includes all the BD and as few as
possible valid data.
datao bae : the aptitude
Cotection oJ _the
for clearing the final data base is of great practical
importance and one of the most essential tasks of the
overall state estimation process.
_na.t
ReoSoniio
o_pagegcay_undentieabte
BD:
Tmptementation_&eqyWLemients
: practical consider-
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3039
3.1. IDENTIFICATION BY ELIMINATION (IBE)
Conceptually, this identification is the continuationof theBD detection step which is a global criterion
implying the residual vector r . The leading idea is
that in the event of a positive detection test, a first
list of candidate BD is drawn up on the basis of an rN
(or rW ) test, then successive cycles of eliminationreestimation-redetection are performed until the detection test becomes negative.
Two subclasses may be distinguished corresponding
to the elimination of single or of grouped BD. Introduced by Schweppe et al. [1] almost at the same time
with the state estimation itself, the former consists
in eliminating at each cycle the measurement having the
largest magnitude of the normalized or weighted residual. As for the grouped elimination, a grouped residual
search has been proposed by Handschin et al. [31; it
consists in eliminating a group of suspected measurements which supposedly includes all BD, and reinserting
them afterwords one-by-one.
Another variant of these procedures consists in
solving eqs. (3) with respect to one or several suspected
measurement errors, then in correcting them by substracting these errors. This measurement error estimation has
first been proposed by Aboytes and Cory [6]. Later on,
Garcia et al. [7,8] have explored the simplified way of
correcting one measurement at a time (the one having at
each step the largest IrNil ) and keeping the W matrix
constant during the subsequent computations of rN. Note
that this technique has also been applied by SimoesCosta et al. [14] to the orthogonal row processing
sequential estimator. The work by Xiang Nian-de et al.
[9-11] has significantly contributed to elucidate this
question. These authors have brought up the singular
character of W , have proposed its partitioning so as
to estimate only s (s< m-n) out of the m measurement
errors. Moreover, they have clearly pointed outthe fact
that correcting these s measurements amownts to eZiminating them. Attempting to improve this technique, MaZhi-
quiang proposed to process combinatorial sets of suspected measurements and to identify the BD through a
detection test based on an interesting formula he established in Ref. [121 (see 4.1. 3 below). Now, becauseof
the equivalence between correction and elimination, the
fact remains that all these techniques belong to the
class of the procedures by elimination.
4.1.1. Description
The methods of this class rely on the rW or the
rN test. The choice between rW and rN implies a tradeoff between good applicability features (simplicity,time
and core savings) and reliability. Generally, the poor
performances of rW (apart from the special caseof high
redundancy and single BD) make the rN test worth-conceding the additional implementation effort. Nevertheless, the latter is not reliable enough either; indeed,
in case of multiple interacting BD, the one-to-one corand erroneous mearespondence between largest
surement stops being guaranteed : valid measurements
may thus be declared false and vice-versa.
Note that the decision is taken on a global basis
given by the sole detection test, which just informs
about the existence of BD among the measurements, but
does not indicate whether the eliminated ones are actually erroneous.
IrN
4.1.2. Assessment
Puos
it is simple, since the only computation it needs besides estimation is that of residuals;
* it is capable to warn the operator that the BD are
topologically unidentifiable, provided the method of
2.4 is implemented.
*
* it is heavy since it requires a series of reestimation-detection after each elimination; this may lead
to computer times incompatible with the on-line requirements;
* it may lead to a degradation of the measurement configuration and a subsequent drop of the power of the
detection test (see fig.1); thisin turnmay cause an
important probability of non-detecting remaining BD
(especially when they become critical);
* it can provoke an undue elimination of valid measurements causing not only a rough identification but
also a drop of the detection test power. When using
the rN test, this situation arises in the case of
multiple interacting BD or of BD located in regions
with low local redundancy, i.e. in the case of stringent identification conditions. On the other hand,
the rw test may lead to a degradation even in mild
situations.
r..
rs = rr5
measurements,
(9)
s-dimensional subvector of
es '
(10)
Here xc is the new state vector obtained from the measurements corrected by es (i.e. eliminated). Therefore,
2
J xc has a x _distribution with (m-n-s) degrees of
freedom.
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3040
The advantages and drawbacks of the above techniques are summarized hereafter.
PU4o
* The main attractiveness of these techniques is that
the correction does not affect the measurement configuration. Hence, the gain matrix can be kept constant during the successive reestimations of the whole
identification procedure, while keeping the goodness
of the minimization procedure convergence. On the contrary, eliminating BD may deteriorate this convergence.
It may even happen that such a procedure which converges properly through the above technique, diverges
when eliminating the BD.
Cons
4.2. IDENTIFICATION BY
NWC
4.2.1. Description
The NQC methodology consists in minimizing the cost
function
m
(1 1)
J (x) =
fi (ri/ai)
i=1
where fi is equal to rl/4G when I rXi <Y ; here rX,
denotes either rW. or rNi and y is a properly chosen
threshold. When I rXi |Y
fi takes one of the following criteria [3]: quadratic-tangent (QT), quadraticlinear (QL), quadratic-square root (QR), quadratic-constant (QC), etc.
Applying the Gauss-Newton algorithm to minimize
(11) gives the following iterative algorithm [3,171
HTP H [x (Q+1)-x(t)] =H
[z -h(x(9))]
(12)
where P and Q are diagonal weighting matrices depending on the residuals. The comparison of eq. (12) withthe
corresponding basic WLS algorithm (P Q=R-1) shows
that the method consists in modifying the weights of the
measurements according to their residuals. Fig. 2 indiQ
cates the variation of the weight Qii with the magnitude of the corresponding residual. As can be seen, the
),the stronger the relarger Ithe Irwi (resp.
jection of the corresponding measurement. This figure
compares also the rejection effect of the various
criteria. In particular, the QC criterion is a borderline case, since it purely eliminates those mesurements
whose residuals are larger than y . Therefore minimizing
the QC criterion amounts to eliminating and/or reinserting measurements at each iteration.
IrNi
4.2.2. Assessment
PUZb. The main advantage of the NQC method lies in its
simplicity. Indeed, on one hand, it can be implemented
through a simple transformation of the basic WLS algo-
2
4
0
6
8
FIG. 2: variation of Qiij vs.
10
rWi /y
rithm; on the other hand, the estimation and identification steps are carried out in a single procedure, which
avoids successive reestimations.
Com6. The method suffers from the following serious
drawbacks.
* Possible existence of local minima : this, however,
can be circumvented by using as a starting point the
result of a WLS estimation.
* Strong tendency to slow convergence or even to divergence: the NQC exhibit a slower convergence than the
corresponding quadratic criterion. This can be explained
as follows
- the shape of the cost function is more intricate;
- the rejection of many measurements may lead to numerically unobservable situations, especially in cases of
poor local redundancy and/or multiple interacting BD.
* High risk of wrong identification. Schematically,
the NQC rely on measurements having small residuals
(with respect to y ) and tend to reject the others. Now,
since there is no one-to-one correspondence between
large residuals and large measurement errors (see 4.1)
it may happen that valid measurements are rejected
whereas erroneous ones are kept. In such a case the estimate is much less reliable than that givenby the quadratic estimation without any BD processing.
* No recognition of topologically unidentifiable BD
situations : in this case, results are unpredictable.
Moreover, the convergence is generally affected, since
the NQC tend to reject too many suspected measurements.
* Partial rejection of BD, except for the QC criterion.
This implies a (vicious) compromise between valid and
BD; the accuracy of the resulting estimate is thus corrupted since it is influenced by wrong information contained in the BD. Inspired by Muller [17], Fig. 2 shows
that the QT criterion is more subjectto,thisdegradation.
4,3. IDENTIFICATION
BY HTI
4.3.1. Description
The HTI method comprises three main steps
(i) At the end of a standard detection test, which presumably has shown presence of BD, the measurements are
i.e. in decreasarranged in decreasing values of
ing suspicion. A list' s of selected among the suspected
oof the meameasurements is drawn up and an estimate
surements error vector es is-computed via eq.(9).
test allows verifying
By means of eq. (10), -the J
IrNil,
18s
(4sijI<
(see
[211)
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3041
E[6si]
(14)
esi
StAategy_q
the
decision
is
taken
with
fixed
type
4.3.2. Assessment
Puz
:
The HTI method is generally able to identify all BD
within a single step (or at worst within two steps).
This is especially true for strategy 6 . Concerning
strategy a experience has shown that, when all the
BD have not been identified by the first test, a second one, performed after a reestimation, is sufficient
to complete the identification. Note that in both
strategies, situations where all BD have not been selected may lead to a slightly larger number of reestimations.
* This method is able to identify strongly interacting
BD. This important advantage results from eq.(14)
which shows that, unlike the residuals, the estimate
esi is not affected by the presence of BD among the
other measurements. In other words, the very notion
of interacting BD becomes meaningless.
* The method treats properly topologically unidentifiable BD. Indeed, the procedure of 2.4 applies to
the HIT method as well.
Conz :
* There isarisk of poor identification, corresponding
to the case where one or several BD are not selected.
This risk can however be alleviated through appropriate techniques [21].
* The method requires the computation of the Wss matrix,
whereas the other procedures merely need the diagonal
of the W matrix. Note however that the technique proposed in [211 avoids necessity of computing the complete E matrix.
*
Ies.I= 40
Hen
,w
also
N= -2.32
4u - 2.jv3lIii_l
and
with
discussed.
of the grouped
elimination, only the single elimination scheme is considered here. However, in order to decrease the number
(NIa)max= 3
.la -1
0< vi< 3
(15)
(15')
questions
1%
tion
5=
hi
FLP
1-2
FLQ 1-2.
INP 1
INQ 1
177.3
-25.7
261.2
-27.1
"Measured" Value e1 = z
z
0.0
30.0
0.0
30.0
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
-hj(x)
-177.3
55.7
-261.2
57.1
e!|
118.2
37.1
174.1
38.1
3042
TABLE II SUCCESSIVE LISTS OF SUSPECTED MEASUREMENTS IN THE SIMPLE ELIMINATION PROCEDURE THROUGH THE rN TEST
lst estimation
Active
FLP
INI)
FLP
FlP
INP
2-1
1
1-3
1-2
2
3rd estimation
2nd estimation
Reactive
rNI JReactive
Active
-81.7
-74.1
49.8
-46.7
INP
FLP
INP
FLP
FLP
rN| Active
2 -71.8 INQ
1-3 56.6 FLQ
1 -40.5 INQ
4-3 -28.7 FLQ
2
29.3 FLP
1-2 15.7 INP
5
13.4 FLP
1-3 -12.0 FLP
1-2 -24.0 FLQ 4-3 10.6 FLP
4th estimation
rNi
rNi Reactive
Active
Reactive
rNi
IrNIl
rN.
5th estimation
<
Active rNN-Reactive
IrNil
IrNil
< 3
J()
rNi
< 3
o n rN
The identification procedure requires four successive elimination-reestimation cycles, after the alarm of
the detection test. They are summarized in Table II. The
elimination of the fourth active measurement makes critical two others. The final list of measurements labelled
false is thus the following.:
- eliminated : FLP 2-1, FLQ 2-1; INP 2, INQ 2; FLP 1-3,
FLQ 4-3; INP 1;
- become critical : FLP 4-3, FLP 1-2.
The final state estimate is the one obtained at the
end of the fifth estimation; some characteristic values
are reported in column four of Table IV (see next page).
The results inspire the following comments.
(i) Both erroneous active measurements are present in
the final list, even if one of them has been included
thanks to the critical measurement analysis.
(ii) Three valid measurements have incorrectly been declared false.
(iii) None of the two erroneous reactive measurements
has been identified. Indeed the improper elimination of
three valid (reactive) data caused an important weakening of the measurement configuration. This in turn provoked a decrease in the value of the Wii coefficients
and hence in the detection capability, as described in
2.2. A more detailed analysis of this question is
given below.
(iv) The final state estimate is completely erroneous
in a certain neighbourhood of node 1, since FLP 1-2,
FLQ 1-2 and INQ i have not been eliminated.
e2
<
X+ NP
(16)
ej +
2 e
<
X+NPd
(17)
I1 = I21
<
-0.380ej + 0.724 e2
< 4.28
(19)
TABLE
III
- SUCCESSIVE VALUES
Before
any elimination
FLQ 1-2
INQ 1
OF
After elim. of
INQ 1
FLQ 1-2
INQ I
FLQ 1-2
0.785
- 0.275
0.5
0564 -.430
0
-.430
0.382
0.524
TO
BD
After elimination of
FLQ 2-1, INQ 2 and FLQ 4-3
FLQ 1-2
.344
-
4.330
I,NQ
0.0330
0.336
P >90%
-4.28 <
-4.28 <
0.587el
-0.569ei
- 0.563 e2 <
4.28
(20)
(21)
5.2.1.2.
EUinaLtion bo6ed
on rw
IrWiJ>
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3043
1st Selection
Selectedr.
measuremnent
FLP
INP
FLP
FLP
FLQ
FLQ
FLP
FLP
INQ
FLP
INQ
FLP
FLQ
FLQ
FLQ
FLP
FLQ
INQ
FLQ
FLP
FLQ
FLP
2-1
1
1-3
1-2
2-1
1-2
4-2
6-2
1
2-6
2
2-5
1-3
4-2
6-2
6-8
6-7
5
2-6
4-6
6-8
6-4
FLQ 2-5
FLQ 4-6
FLP 6-9
QT
QL
OR
Susp. measurts.
rWl
1
1-2
2-1
1-3
2
6-2
2-5
-103.9
-11.8
10.0
-4.4
-3.2
3.1
INP 1
FLP 1-2
FLP 1-3
FLP 2-1
-162.6
-111.0
5.9
-5.2
INP
FLP
FLP
FLP
INP
FLP
FLP
INP 1
FLP 1-2
-151.4
-171.4
-116.4
Susp. measurts.
FLQ
INQ
FLQ
INQ
FLQ
FLQ
1-2
1
2-1
2
esi
esi
1i
r11
1i
xi
2.30
-261.20
2.33
-177.30
-2.86
55.69
1.09
-0.64
57.06
-1.83
-0.78
1.61
-2.67
-23.27
-211.65
23.51
-148.89
19.34
41.84
-9.67
-8.79
39.74
7.03
17.63
5.92
-6.15
4.95
2.94
1.37
-6.72
-6.31
-2.14
-10.12
-0.18
9.22
0.38
1.25
1.02
97.47
173.51
73.86
99.75
44.75
39.39
41.94
34.37
53.51
35.64
42.99
19.70
16.12
15.15
15.54
11.76
21.06
22.36
12.32
29.01
11.71
28.50
1056.00
3345.00
606.20
1106.00
222.50
172.40
195.40
131.20
318.10
141.10
205.30
43.12
28.89
25.50
26.83
15.37
49.30
55.55
16.88
93.52
15.24
90.23
7.68
2.31
2.08
0.00
0.00
0.00
0.00
0.37
0.73
0.55
1.18
0.00
1.06
0.48
3.00
3.00
3.00
3.00
3.00
3.00
3.00
3.00
1.83
3.00
1.83
3.00
3.00
3.00
0.00
O.00
0.00
0.00
0.23
-1.59
1.29
0.70
1.82
0.30
-0.30
-0.79
-0.42
-0.01
1.31
1.22
8.32
4.56
4.33
8.28
14.38
11.53
20.28
0.00
1T. 89
10.32
29.55
24.19
22.72
23.31
17.64
31.60
33.54
18.49
26.55
17.57
26.07
12.48
6.84
6.49
TABLE VII
rW_
1-3
4-2
7.9
-7.7
3.7
FLQ 1-2
INQ 1
FLQ 2-1
INQ 2
FLQ 1-3
26.9
23.7
9.8
7.7
-5.9
FLQ 1-2
INQ 1
FLQ 1-3
33.8
33.2
-3.3
eSi
4.47
-261.23
-177.28
4.91
54.32
57.42
4.91
6.64
4.98
20.72
20.68
-0.94
1.37
7.04
3.84
Selected measurements
FLP
INP
FLP
FLQ
FLQ
INQ
INQ
FLQ
FLQ
FLQ
-FLQ
22.1
19.1
13.8
rN test, only 25
Strategy 8
Str. a
2-1
1
1-2
2-1
1-2
1
2
1-3
4-2
6-7
2-6
7.86
26.63
23.30
-3.57
5.46
0.32
3.78
TABLE VIII
Strategy B: 2nd selection
Select.
Meas.
es
Xi
vi
rj
'i~i~
M
FLP
INP
FLP
FLP
FLQ
FLQ
INQ
INQ
FLP
INP
FLQ
ds j
rij1 vi1
Xi
1
INP 1 -257.25
FLP 1-2 -174.51
FLQ 1-2 60.34
64.91
INQ 1
performed.
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3044
TABLE IX
Actual
value
"Measured"
value
82.6
2.8
17.6
7.0
-0.5
2.5
-2.4
184.6
101.7
69.2
56.1
19.0
22.4
-12.1
-10.2
hi(x)
2-5
2-5
12-15
12-15
24-25
24-25
29
INQ 29
FLP
FLQ
FLP
FLQ
FLP
FLQ
INP
Zi
-0.9
ei
68.0
65.9
64.5
61.4
24.4
24.9
12.1
11.6
-9.3
0.84
0.85
0.15
0.16
0.62
0.64
0.47
0.47
Tables VI and VIII summarize the results of strategy 6 . Four cycles of selection were needed. The six
suspected measurements which were not inserted in the
first selection (for observability reasons) are introduced in the second and third ones. Note that for the
first test, the value of 'i is equal to zero for five
measurements : this results from the poor accuracy of
the correspondinq estimates. However, for most of the
measurements, Vi reaches its maximal value (3.0) atthe
second test. This shows the rapid increase in-accuracy
of the estimates and hence in power of the identification test. Finally, the fourth selection is simply composed of the four BD.
others).
5.3.1. Identification by elimination
5.3.1.1. IBE bazed on rN
The procedure has required 5 successive cycles corresponding to the following final list :
- eliminated : INP 5, FLQ 2-5; FLP 2-5, FLQ 12-15; FLP
12-15, FLQ 24-25; FLP 24-25, INQ 29; INP 29;
- become critical : INP 2.
All the BD have been eliminated. The incorrect elimination of INP 5 has made INP 2 critical. Note that the
latter measurement is not erroneous; however this cannot
be verified a posteriori.
5.3.1.2. IBE baed on rw
The identification has required 7 successive reestimations. The final list of measurements declaredfalse
is the following.:
- eliminated : FLP&FLQ 2-5; FLP&FLQ 24-25; FLP 12-14,
FLQ 12-15; FLP 12-16, INQ 29; FLP&FLQ 4-12; FLP10-17;
MODV 13; INP 29, MODV 12;
- become critical : FLP 12-15, INP 16, INP 17.
Measurements Actual BD
labelled
false
Valid data
Qualitytof
state estimation
NNumber
uberof
of
4 Interacting BD
IBE
8 Noninteracting BD
'W rN
QT
NQC
QL
QR
rw
of
[state
[T; Number
reestimations | 4
METHOD
PERFORMANCE
CRITERIA
TABLE X
Conclusions are similar to those drawn for the preceding case, even if the identification conditions are
less stringent here. As in the interacting case, the QC
has been unable to provide an estimation. Note that,
because of a low local redundancy, the quality of the
state estimation in the vicinity of node 15 is rather
bad for all the NQC (see Table X).
It is worth-mentioning that the NQC efficiency
is found to vary with the noise attached to the
valid measurements. This gives NQC a "capricious"
behaviour.
1 |
___
2
2 __
23
HTI
cr iterations/estimation
Time in sec. CPU
1.8 2.5 5.0 5.5
m
Number of
455 Q the Z matrix terms
to be computed
:___.
__5
1.1
NQC
HTI
rN
QT
QL
QR
15
ad
go
godbd
3 .__
3
1.9
1.4
560
560
2
_ 24
IBE
3.2
3.1
_____
8
8
_ 10 _
3_
__
3
1.7
2.2
1.5
1.7
1
1.7
455
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
1360 1360
3045
6. CONCLUSION
The identification techniques available today have
been classified into three broad classes; their capabilityto face various types of BD has been found to differ
significantly from one class to another.
The NQC exhibit the most poor performances; they
are very sensitive to low local redundancy and to interaction of BD; they have a slow convergence and a "vicious".behaviour. In brief, they don't show to be suitable enough.
On the other hand, the IBE techniques are attractive with respect to implementation considerations:
they are easy to use and simple to implement. They show
to be quite interesting as long as the BD are non- (or
weakly) interacting and located in regions of moderate
redundancies. They start being unefficient, however,
when the number of BD and their spreading increase and
when the local redundancy decreases. Although much more
reliable than the NQC, the IBE methods lead to inaccurate BD identification results at a certain level of
severity of the identifiability conditions.
The HTI method, finally, seems to combine effectiveness, reliability and compatibility with on-line
implementation requirements. This latter aspect receives
at present further consideration.
REFERENCES
[1] F.C. Schweppe, J. Wildes, D.B. Rom, "Power System
Static State Estimation. Parts I, II, III", IEEE
Trans. on PAS, vol.PAS-89, No.1, Jan.1970, pp.120-135.
[21 J.F. Dopazo, O.A. Klitin, A.M. Sasson, "State Estimation for Power Systems : Detection and Identification of Gross Measurement Errors", Proc. of the 8th
PICA Conf., Minneapolis, 1973, pp. 313-318.
[3] E. Handschin, F.C. Schweppe, J. Kohlas, A. Fiechter,
"Bad Data Analysis for Power System State Estimation", IEEE Trans. on PAS, vol.PAS-94, No.2, March/
April 1975, pp. 329-337.
[41 A. Merlin, F. Broussole, "Fast Method for Bad Data
Identification in Power System State Estimation",
Proc. of the IFAC Symp., Melbourne, Feb.1977, pp.
449-453.
[5] N.Q. Le, H.R. Outhred, "Identification and Elimination of Bad Data and Line Errors for Power System
State Estimators", Proc. of the IFAC Symp., Melbourne,
Feb.1977, pp.459-463.
[7]
[8]
[91
[10]
[11]
[12]
[131
APPENDIX
The IEEE 30-bus system, along with the measurement
configuration is schematically given inthe figurebelow.
It comprises 118 measurements leading to a redunaancy
T= 2 . The following standard deviations have been used
- for power measurements : 0= 1.5 MW/MVAr at 132 kV and
C= 0.8 MW/MVAr at 33 kV;
- for voltage measurements: 0= 0.005 p.u.;
- for injection pseudo-measurements: a= 0.2 MW/MVAr.
Detection and Identification Techniques Using Estimation Orthogonal Methods", IEEE Trans. on PAS, vol.
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3046
Discussion
M. S. Kurzyn (Transmission Development Department, State Electricity Commission of Victoria, Melbourne, Australia): The HTI method uses
ranking of suspect measurements and one-shot identification procedure,
both being the salient features of the simple bad data identification scheme
described in [A]. However, many details of this method are necessarily
different from those of [A], and the authors should be commended for
a development of what appears to be a highly effective bad data identification technique.
The HTI method and two representatives of the existing methods have
been tested and compared using four networks, but perhaps due to space
limitations the paper shows only the results pertaining to the smallest
network. Could the authors bother to present the remaining test results?
Is the HTI method effective and reliable for the larger networks as well?
What is the sensitivity of the HTI method to different system operating
conditions?
The authors' response to the above questions would be greatly
appreciated.
REFERENCE
[A]. The above reasoning implies that the HTI may fail to identify "correctly" the bad data. Have the authors encountered failed cases in their
testing of HTI method?
It would be very helpful if the authors could give explicitly step-bystep of the HTI algorithm. Though not explicitly stated stated in the
description of the HTI method, the testing of observability in selecting
suspected measurements is an important part in the algorithm. Would
the authors care to comment on (i) what method is used for observability test, (ii) what is the percentage of computation time spent on it, and
(iii) the effect of decreasing in measurement redundancy on the observability test.
REFERENCES
[A] A. Monticelli, F. F. Wu, and M. Yen, "Multiple Bad Data Identification for State Estimation by Combinatorial Optimization,"
the IEEE Power Industry Computer Application Conference,
(PICA) pp, 452-460, May 6-10, 1985, San Francisco.
[B] R. J. Kaye, "A Geometric Approach to Bad Data Analysis in Electric Power Systems State Estimation," to be presented at 1985 International Symp. on Circuits and Systems, June 5-7, 1985, Kyoto.
Manuscript received March 1, 1985
L. Mili, Th. Van Cutsem, and M. Ribbens-Pavella: We thank the
discussers for their interest in our paper and their constructive remarks.
We shall group our answers by subject matter, starting with those
relative to IBE, then to HTI method; for the latter, increasing order of
generality will be followed.
Professors Wu and Monticelli discuss many interesting issues relative
to IBE and HTI methods.
A. As concerning the IBE, they raise two practical aspects considered
but not developed enough in the paper because of space limitations. We
are clarifying them hereafter.
A.1 With regard to the frequency of failure of the IBE method, we
don't share the discussers' opinion that multiple interacting bad data cases
where IBE fails are "special and artificial". Of course, the example of
Section 5 of the paper was chosen for the purpose of illustrating the
theoretical considerations of previous sections. But we have never claimed
that IBE performs always as unsatisfactorily as it does in this example.
Nevertheless, it is not true either that these cases are "rare." A means
to tentatively assess IBE's frequency of failure is given below.
Let us consider the IEEE-30 bus system of the paper with the two erroneous active measurements :FLP1-2 and INP1 (similar results are obtained in the reactive case). Let us identify the domain D of the corresponding measurement errors (ek,ee) for which the IBE method unduly
eliminates valid measurements and fails to identify the two bad ones.
This domain is such that the normalized residual of a bad data is never
the largest one, i.e.
D= U di with
i valid
i k,f
and
IrNel <IrNilI
WitZ
Wik
e
,(Al)
.oi.,/w
ork VWkk
0k \/Wkk
,Wi i
OFi
A similar relation can be written for the inequality rN I < IrNi (i * ).
Substituting the numerical values of the Wij terms, the corresponding
domain Di can be easily determined. Fig. A shows the final domain D
(note that in this particular case, the domain Di corresponding to measurement FLP 2-1 contains all the other Di). Assuming that the errors ek
and ei are bounded by measurement full scale values, we obtain a ratio
I Wkk
Wkk
<
of failure of about 10%. Taking into account the effect of the noises
in (Al) results in a negligible decrease in this ratio (see dotted lines on
Fig. A). Note that after the undue elimination of measurement FLP 2-1,
the new corresponding domain D includes the preceding one, so that
another valid measurement (INP 2) will be elmiinated. Therefore we may
conclude that the domain D of Fig. A is the domain of failure of the
IBE method, for the two bad data.
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3047
Diag(W) refreshed
INP
FLP
FLP
FLP
FLP
FLP
FLQ
FLP
INQ
FLP
IVf
FLP
FLQ
FLP
FLP
29
30
29-27
27-29
26
27-28
27-30
24-25
28- 6
6-28
24-25
24-22
26
28- 8
27
4-12
4-12
6-9
12-15
rN
118.95
108.46
- 45.10
42.84
34.22
-26.07
21.63
13.56
-9.88
8.83
-6.09
5.67
-4.15
-3.95
-3.87
-3.64
-3.22
-3.20
-3.19
COLUMN II:
2nd estimation
Measurts.
rM
COLUMN III :
3rd estimation
Measurts.
rN
COLUMN IV:
2nd estimation
Measurts.
rN
COLUMN V
3rd estimation
Measurts.
rN
performed
performed
correction
correction
INP 29: -42.69 MW FLP27-30 :- 14.9 MlW
performed
correction
INP 30 : 23.20 MW
adequate correction
INP 29 : - 5.52 MW
INP 30 :
10. 19 MW
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3048
conclusion that the W- lss matrix - required in the HTI method - is
anyway necessary to properly and efficiently face the problem of multiple bad data, whatever their interaction.
Wji
initial
INP 29
configuration eliminated
Measurements
0.445
0.352
0.773
0.814
0.354
0.714
0.840
0.610
0.944
0.469
INP
INP
FLP
FLP
INP
FLP
FLP
FLP
FLP
29
30
29-27
27-29
26
27-28
27-30
24-25
28- 6
INQ 29
INP 29
and
INP 30
eliminated
0.157
0.665
0.629
0.303
0.698
0.840
0.608
0.940
0.465
0.574
0.557
0.260
0.670
0.262
0.602
0.938
0.469
[21]:
(A2)
es = W-Issrs = Zs - hs(xc)
where xc is the state estimate based on the remaining zt measurements,
i.e. those remaining after eliminating the s selected ones.
This property is precisely at the root of the equivalence between correction and elimination (as was already observed by the authors of Ref.
[11]). However, the discussers do not consider the counterpart of the
elimination: eliminating valid measurements wastes useful information,
decreases the power of the detection test, and increases the interaction
of the remaining bad data. These drawbacks are inexistent in the HTI
method which performs on the initial configuration: Eq. (A2) shows that
es is based on Zs as well as zt.
B.2. Considering projection properties is indeed worthwhile since it
allows gaining good insight into the theoretical understanding of the WLS
estimator, as has already been observed and used in Ref. [C] below.
Within this context, we agree with the discussers that the residual vector
r belongs to N(HTR - 1), since
(A3)
HR-TRlr=O
Now, what probably has escaped the discussers' attention is that the vector
of concern in POE and hence in HTI is not r but a subvector rs of it,
for which the above property is no longer valid. Indeed, partitioning (A3)
according to selected and other measurements yields:
T -1
T -1
HSRSr5+ H~Rt
rt
Wsses
Wstet
Among all the possible solutions, HTI definitely provides the OPTIMAL
one in the statistical sense, since the estimate Rs = W- 'ssrs is:
(i) unbiased, provided that all erroneous measurements are selected, which
is ensured by adequate measurement selection (see Ref. [21]);
(ii) minimum-variance (i.e. es is the most accurate estimate of es): this
important property is a recent result, shown and discussed in Ref. fD,E].
with
-1
w-l
-1
HTR-1=I
(I- H
+H
T-1
ZHJR
=
5+
I~St=XZ15[R-15H]'4
HS 5[Rs H ZHS] t ss
(A5)
(A6)
These two formulae are used in a recursive way, each time a new measurement is added to the selection; in fact, in this case, Eq. (A6) reduces to:
where
n-s
D1
t
2
1
a-hi
+n-s+I Ii lYkT-s+i
nm-s+1
hi
(scalar)
(A7)
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.
3049
Fig. B. The overall algorithm
Make a (proper) initial selection
[measurements are classified into: selected)
stand-by} (suspected)
others (valid)
REPEAT:
Apply the hypothesis test to each selected measurement
If the test is positive for some measurements:
THEN: make a new selection =
measurements with positive test
+ as many stand-by measurements as possible;
UNTIL the test is negative for all the selected measurements;
[the latter are the bad data]
correct selected measurements : Zs = zs - es
re-estimate the state (with fixed gain matrix).
s =0
F. = E
REPEAT:
Search for the (not yet selected) measurement with
maximum rNi ; let i be this measurement
compute
Di = 2ihi
rss Is fs HSR
s
eS-=rS Srs-
4..-
2
UNTIL s= smax or Jc <Xm-n-s
FOR each suspected, neither already selected nor already in
stand-by measurement
IF this measurement had initially a Wii > EW
THEN : compute Di = - hTE'"h
IF Di <eDoal
THEN : this measurement is also in stand--by
2. The only difference existing between small and large networks concerns the computation time. As for the HTI method, the difficulty (particularly for large systems) presented by the computation of Ws5 may
be cleared through efficient ways: (i) The submatrix of L involved in
W5s may be calculated through an updating of the factors of G. Presently,
this new approach is under testing. (ii) A straightforward method is now
available. It consists of multiplying the inverse factored form of G by
the appropriate columns of the identity matrix and of using the sparse
vector technique (see Ref. [F]). As mentioned in the closure of this latter paper, this technique speeds up dramatically the computation of the
off-diagonal terms of the inverse matrix. Let us mention that, whatever
the size of the network, a selection involving more than 30 measurements
is rare.
3. About the sensitivity of HTI method: experience has shown that
the linear assumption is valid whatever the operating point of the system
and whatever the measurement weighting factors.
Finally, we would like to draw some general conclusions inspired by
the experience we gained during the implementation of HTI.
(i) In obvious single bad data cases, the HTI method reduces to
the conventional IBE method.
(ii) In the case of multiple (whatever interacting or not) bad data
the main part of the computing effort, i.e., computation of
W- lss matrix is required of both, IBE and HTI methods. At
the expense of a very little additional effort, HTI guarantees
the identification of all bad data, and only of them.
(iii) The HTI method is easy to program and to implement (we have
implemented it on a PDP-1 1/70 computer which is in charge
of various tasks currently existing in the control center). Of
course, a prerequisite of its programming is its understanding;
in other words, one cannot expect an inexperienced engineer
(who, e.g., has seldom got involved with state estimation-related
functions) to successfully program it.
(iv) The claim according to which one should not bother about
special multiple interacting bad data cases, which are anyway
likely to be successfully handled by the system's operator is in
our opinion arguable. Indeed, as long as the system is small
and the operator experienced, the need for fully automatic state
estimation, security and the like functions is hardly felt. The
difficulties arise when complexity increases beyond human
capacities.
REFERENCES
[C] K. A. Clements, G. R. Krumpholz, P. W. Davis, "Power System
State Estirriation Residual Analysis: An Algorithm Using Network
Topology", IEEE Trans. on PAS, vol. PAS-100, no. 4, pp.
1779-1787, April 1981.
[DI L. Mili, Th. Van Cutsem, M. Ribbens-Pavella, "Decision Theory
Applied to Bad Data Identification in Power System State Estimation", to be presented at the 7th IFAC Symp. on Identification
and System Parameter Estimation, Univ. of York, U.K., July 3-7,
1985.
El L. Mili, "Algorithmes Fiables d'Identification des Fausses Donnees par Tests d'Hypotheses", Internal Report (in French), Univ.
of Liege, No. MBC/3, May 1984.
[F] W. F. Tinney, V. Brandwajn, S. M. Chan, "Sparse Vector
Methods", IEEE Trans. on PAS, vol. PAS-104, no. 2, pp. 295-301,
Feb. 1985.
Authorized licensed use limited to: to IEEExplore provided by Virginia Tech Libraries. Downloaded on January 16, 2010 at 13:31 from IEEE Xplore. Restrictions apply.