H Y , ... , H
(
Y , ... , H Y Pr
)
k
m ,
k
(r
)
j
m ,
j
r
)
1
m ,
1
(r
-
=
-
=
-
=
j
j
1
1
1
k
r
0
k
i
1
j
r
0
j
i
1
1
r
0
1
i
i
m
...
i
m
... ... =
D
o
w
n
l
o
a
d
e
d
b
y
[
R
y
e
r
s
o
n
U
n
i
v
e
r
s
i
t
y
]
a
t
1
0
:
4
6
0
5
J
u
n
e
2
0
1
3
N.A. Nechval, K.N. Nechval, E.K. Vasermanis / AVIATION 2004, Vol VIII, No 3, 39
- 5 -
( ) ( )
,
i
m
H Y Pr H Y Pr
i
m
...
) m , (i ) m , 1 (i
k
k
S
S
S S S
+
S
(1)
where
=
S S
k
1 j
j
k
1 = j
j
. m = m , i = i (2)
Proof.
H Y , ... , H Y , ... , H Y Pr
)
k
m ,
k
(r )
j
m ,
j
(r )
1
m ,
1
(r
=
H Y Pr =
)
j
m ,
j
(r
k
1 j
[ ] [ ]
-
-
= =
j
i
j
m
j
i
1
j
r
0
j
i j
j
k
1 j
) H ( F 1 F(H)
i
m
E =
-
=
-
=
-
=
j
j
1
1
1
k
r
0
k
i
1
j
r
0
j
i
1
1
r
0
1
i
i
m
...
i
m
... ... =
[ ] [ ] { }
S
-
S
- S
i m i
k
k
) H ( F 1 F(H) E
i
m
... . (3)
Since
[ ] [ ] { }
S
-
S
- S
i m i
) H ( F 1 F(H) E
[ ] [ ]
[ ] [ ]
-
-
=
-
S
-
S
=
S
-
S
S
=
S
-
S
S
i m i
1 i
0 i
i m i
i
0 i
1
) H ( F 1 F(H)
i
m
) H ( F 1 F(H)
i
m
E
i
m
( ) ( )
,
i
m
H Y Pr H Y Pr
=
) m , (i ) m , 1 (i
-
S
S
S S S
+
S
(4)
The joint probability can be written as
H Y , ... , H Y , ... , H Y Pr
)
k
m ,
k
(r )
j
m ,
j
(r )
1
m ,
1
(r
-
=
-
=
-
=
j
j
1
1
1
k
r
0
k
i
1
j
r
0
j
i
1
1
r
0
1
i
i
m
...
i
m
... ... =
( ) ( )
.
i
m
H Y Pr H Y Pr
i
m
..
) m , (i ) m , 1 (i
k
k
.
S
S
S S S
+
S
(5)
This ends the proof.
Corollary 1. If r
j
= 1, "j=1(1) k, then
H Y , ... H, Y , ... H, Y Pr
)
k
m (1, )
j
m (1, )
1
m (1,
( ) H Y Pr =
) m (1,
S
. (6)
2. Invariant embedding technique for
obtaining prediction limits
This paper is concerned with the implications of
group theoretic structure for invariant performance
indexes. We present an invariant embedding technique
based on the constructive use of the invariance principle
of mathematical statistics. This technique allows one to
solve many problems of the theory of statistical inferences
in a simple way. The aim of the present paper is to show
how the invariance principle may be employed in the
particular case of finding prediction limits. The technique
used here is a special case of more general considerations
applicable whenever the statistical problem is invariant
under a group of transformations, which acts transitively
on the parameter space.
2.1. Preliminaries
Our underlying structure consists of a class of
probability models (X, A, P), a one-one mapping y taking
P onto an index set Q, a measurable space of actions (U,
B), and a real-valued function r defined on Q U. We
assume that a group G of one-one A - measurable
transformations acts on X and that it leaves the class of
models (X, A, P) invariant. We further assume that
homomorphic images G and G
~
of G act on Q and U
respectively. ( G may be induced on Q through y; G
~
may
be induced on U through r). We shall say that r is
invariant if for every (q, u) Q U
), u , ( r ) u g
~
, g ( r q = q gG. (7)
Given the structure described above there are
aesthetic and sometimes admissibility grounds for
restricting attention to decision rules j: X U, which
are (G, G
~
) equivariant in the sense that
G. g , x (x), g
~
(gx) j = j X (8)
If G is trivial and (7), (8) hold, we say j is G-
invariant, or simply invariant [5].
2.2. Invariant Functions
We begin by noting that r is invariant in the sense of
(9) if and only if r is a G
is
defined on Q U as follows: to each gG, with
homomorphic images g
~
, g in G
~
, G respectively, let g
(q,
u)= u) g
~
, g ( q , (q, u)(Q U ). It is assumed that G
~
is a
homomorphic image of G .
Definition 1 (Transitivity). A transformation group
G acting on a set Q is called (uniquely) transitive if for
every q, JQ there exists (unique) G g such that
g q=J.
D
o
w
n
l
o
a
d
e
d
b
y
[
R
y
e
r
s
o
n
U
n
i
v
e
r
s
i
t
y
]
a
t
1
0
:
4
6
0
5
J
u
n
e
2
0
1
3
N.A. Nechval, K.N. Nechval, E.K. Vasermanis / AVIATION 2004, Vol VIII, No 3, 39
- 6 -
When G is transitive on Q we may index G by Q:
fix an arbitrary point qQ and define
1
g
q
to be the
unique G g satisfying g q=q
1
. The identity of G
clearly corresponds to q. An immediate consequence is
Lemma 1.
Lemma 1 (Transformation). Let G be transitive on
Q. Fix qQ and define
1
g
q
as above. Then
1
q
g
q
=
1
g q
q
for
qQ, G q .
Proof. The identity q = q = q
q q
1 1
g q q g
1 q
shows that
1
q
g
q
and
1
g q
q
both take q into
1
qq , and the lemma
follows by unique transitivity.
Theorem 2 (Maximal Invariant). Let G be
transitive on Q. Fix a reference point q
0
Q and index G
by Q. A maximal invariant M with respect to G
acting on
Q U is defined by
. u) , ( , u g
~
) u , ( M
1
U Q q = q
-
q
(9)
Proof. For each (q,u)(Q U ) and G g
u g
~
) g
~
g
~
( u g
~
) g
~
( ) u g
~
, g ( M
1 1
g
-
q
-
q
= = q
) u , ( M u g
~
u g
~
g
~
g
~ 1 1 1
q = = =
-
q
- -
q
(10)
by Lemma 1 and the structure preserving properties
of homomorphisms. Thus M is G
(q
2
,u
2
) for some g
, and the
proof is complete.
Corollary 2.1 (Invariant Embedding). An invariant
function, r(q,u), can be transformed as follows:
), , v ( r ) u g
~
, g ( r ) u , ( r
1
h = q = q
-
q
-
q
& & (11)
where v=v(q, q
)
) is a function (a pivotal quantity)
such that the distribution of v does not depend on q;
h=h(u, q
)
) is an ancillary factor; q
)
is the maximum
likelihood estimator of q (or the sufficient statistic for q).
Corollary 2.2 (Best Invariant Decision Rule). If
r(q,u) is an invariant loss function, the best invariant
decision rule is given by
), , ( u ) x (
1
q h h = = j
* - * *
)
(12)
where
{ }. ) , v ( r E inf arg
v
h = h
h
*
& & (13)
Corollary 2.3 (Risk). A risk function (performance
index)
{ } { } ) , v ( r E )) x ( , ( r E )) x ( , ( R
v x o o
& &
o
h = j q = j q (14)
is constant on orbits when an invariant decision rule
j(x) is used, where ) x , ( v v q =
o o
is a function whose
distribution does not depend on q; ) x , u (
o o
h = h is an
ancillary factor.
For instance, consider the problem of estimating the
location-scale parameter of a distribution belonging to a
family generated by a continuous cdf F: P ={P
q
: F ((x-
m)/s), xR, qQ}, Q={(m,s): m,sR, s>0}=U. The
group G of location and scale changes leaves the class of
models invariant. Since G induced on Q by P
q
q is
uniquely transitive, we may apply Theorem 2 and obtain
invariant loss functions of the form
] / ) x ( , / ) ) x ( [( r )) x ( , ( r
2 1
s j s m - j = j q , (15)
if q=(m,s) and j(x)=(j
1
(x), j
2
(x)). Let ) , (
2 1
q q = q
) ) )
,
u=(u
1
,u
2
), then
), v , v v ( r ) , v ( r ) u , ( r
2 2 2 1 1
h h + = h = q & & & & (16)
where v=(v
1
,v
2
), v
1
= s m - q / ) (
1
)
, v
2
= s q /
2
)
;
h=(h
1
,h
2
), h
1
=
2 1 1
/ ) u ( q q -
) )
, h
2
=
2 2
/ u q
)
.
The invariant embedding technique, which is used
for constructing lower simultaneous tolerance limits, is
based on the result of Corollary 2.1.
3. Examples
Example 1. For instance, suppose that X
1
, ... X
n
and
Y1j, ... Ymj (j=1, ... k) denote n+km independent and
identically distributed random variables from a left-
truncated Weibull distribution with pdf
( ) [ ] 0, , x , x exp x = ) , b a f(x;
1
> d s m s m - -
s
d
d
d d - d
, , (17)
which is characterized by being three-parameter
(m,s,d) where d is termed the shape parameter, s is the
scale parameter, and m is the truncation parameter
interpreted as the minimum time to crack initiation
(warranty period). It is assumed that the parameter d is
known. Let X
(1)
be the smallest observation in the initial
sample of size n and
). X X ( T
) 1 (
n
1 i
i n
d
=
d
- =
(18)
It can be justified by using the factorization theorem
that (X
(1)
,T
n
) is a sufficient statistic for (m,s). Let
) m , 1 (
j
Y be
the smallest observation in the jth future sample of size
m
j
=m, "j=1(1) k. We wish, on the basis of a sufficient
statistic (X
(1)
,T
n
) for (m,s), to construct simultaneous one-
sided lower 100g% prediction limits for , Y
) m , 1 (
j
j=1, ...k. It
follows from Corollary 1 that this problem reduces to the
problem of constructing a lower 100g% prediction limit,
L
g
, for
. Y min = Y
) m (1,
k j 1
km) (1,
j
(19)
By using the above technique of invariant
embedding of (X
(1)
,T
n
), if X
(1)
<Y
(1,km)
, or (Y
(1,km)
,T
n
), if
D
o
w
n
l
o
a
d
e
d
b
y
[
R
y
e
r
s
o
n
U
n
i
v
e
r
s
i
t
y
]
a
t
1
0
:
4
6
0
5
J
u
n
e
2
0
1
3
N.A. Nechval, K.N. Nechval, E.K. Vasermanis / AVIATION 2004, Vol VIII, No 3, 39
- 7 -
X
(1)
Y
(1,km)
, into a pivotal quantity s m -
d d
/ ) Y (
) km , 1 (
or
s m -
d d
/ ) X (
) 1 (
, respectively, we obtain an ancillary statistic
( )
n (1) km) (1,
T X Y = W
d d
- (20)
whose distribution does not depend on any unknown
parameter. The pdf of W is given by
=
0. w if ,
nw 1
1
km + n
1)km - n(n
0, > w if ,
kmw + 1
1
km + n
1)km - n(n
) w ( g
n
n
(21)
Therefore, in this case L
g
can be found explicitly as
( )
( )( )
+ g -
- +
g >
+ g
+
=
d
-
d
d
-
d
g
.
km + n
n
if ,
km n 1
km
1
n
T
X
,
km + n
n
if , 1
km n
n
km
T
X
L
1/
1 n
1
n
(1)
1/
1 n
1
n
(1)
(22)
If, for instance, n=10, d=8, k=3, m=5, g=0.95, X
(1)
=5 (in
number of 10
4
flight-hours), and T
n
=10917240. Then we
find from (22) that, with n/(n+km) = 10/(10+15) < g,
( )( )
km n 1
km
1
n
T
X = L
1/
1 n
1
n
(1)
d
-
d
g
+ g -
- +
( )( )
3.8 =
15 10 0.05
15
1
10
10917240
+ 5 =
1/8
9
1
8
+
- (23)
and we have 95% assurance that no cracks will
occur in aircraft structural components before L
g
=3.8
(10
4
) flight-hours.
Example 2. Let X
(1)
< X
(2)
< ... < X
(r)
be the first r
ordered observations of time to crack initiation for
identical structural components of aircraft from a sample
of size n from a two-parameter Weibull distribution with
probability density function
s -
s s
d
= d s
d
- d
otherwise, 0,
0, x ], ) / x ( exp[
x
) , ; x ( f
1
(24)
where the parameters s and d (s>0, d>0) are
unknown. Two types of censoring are generally
recognized. In Type I censoring, the time, when censoring
occurs, is fixed, and the number of survivors at this time
are random variables. In Type II censoring, which is of
primary interest here, the number of survivors is fixed and
X
(r)
is a random variable. In Type II censoring, the
likelihood may be written as follows:
s
-
s
d
=
d
=
- d
r
1 i
) i (
r
1 i
1
) i (
r
x
exp
x
L
r n
x
1
) r (
dx ] ) / x ( exp[
x
-
d
- d
s -
s s
d
, x ) r n ( x
1
exp x
r
1 i
) r ( ) i (
r
1 i
1
) i (
r
- +
s
-
s
d
=
d d
d
=
- d
d
(25)
=
- d + s d - d + =
r
1 i
) i (
x ln ) 1 ( ) ln (ln r constant L ln
.
x ) r n ( x
r
1 i
) r ( ) i (
d
=
d d
s
- +
-
(26)
This leads to the likelihood equations
, 0
x ) r n ( x
r L ln
1
2
r
1 i
) r ( ) i (
= ds
s
- +
+
s
d
- =
s
- d
d
=
d d
(27)
=
+ s -
d
=
d
r
1 i
) i (
x ln ln r
r L ln
d - d
=
d
s
- + -
) r ( ) r (
r
1 i
) i ( ) i (
x ln x ) r n ( x ln x
. 0 ln x ) r n ( x
r
1 i
) r ( ) i (
= s s
- + +
d -
=
d d
(28)
Then the MLEs s
)
and d
)
are solutions of
,
r
x ) r n ( x
/ 1
r
1 i
) r ( ) i (
d
=
d d
- +
= s
)
) )
)
(29)
.
x ln
r
1
x ) r n ( x
x ln x ) r n ( x ln x
1
r
1 i
) i (
1
r
1 i
) r ( ) i (
) r ( ) r (
r
1 i
) i ( ) i (
-
=
-
=
d d
d
=
d
- +
- +
= d
) )
) )
)
(30)
The results given here apply to the Weibull
distribution in the form (24). The results are presented
more naturally, however, if we consider the variable lnX,
which follows the extreme-value distribution,
, x ln
,
b
a x ln
exp exp
b
a x ln
exp
b
1
) b , a ; x (ln f
< < -
-
-
-
=
(31)
D
o
w
n
l
o
a
d
e
d
b
y
[
R
y
e
r
s
o
n
U
n
i
v
e
r
s
i
t
y
]
a
t
1
0
:
4
6
0
5
J
u
n
e
2
0
1
3
N.A. Nechval, K.N. Nechval, E.K. Vasermanis / AVIATION 2004, Vol VIII, No 3, 39
- 8 -
where a = lns and b =d
-1
. Now (31) is a distribution
with location and scale parameters a and b, and it is well
known that if , a
)
b
)
are maximum likelihood estimates for
a, b from a complete sample of size n, then
, b / ) a a ( -
)
b / ) a a (
)
)
- and b / b
)
are quantities whose
distributions depend only on n.
We are interested in estimating Y
(1,km)
, the smallest
order statistic in all k future samples, each consisting of m
units from the distribution (24). It is easily shown that
,
Y
ln ) ln Y (ln b / ) a Y (ln W
) 1 (
) 1 ( ) 1 (
s
d = s - d = - =
)
)
)
) )
)
(32)
where Y
(1)
Y
(1,km)
, is parameter-free, with distribution
depending only on n and km. Hence, probability
statements for V lead to confidence interval statements for
Y
(1)
.
Let X
(1)
, X
(2)
, ... X
(n)
and Y
(1)
, Y
(2)
, ... Y
(km)
represent
ordered observations. In particular, let X
(1)
< X
(2)
< <
X
(r)
be the first r ordered observations from a sample of
size n from the distribution (24), i.e. we deal with Type II
censoring. The joint density of lnX
(1)
lnX
(r)
is
) b , a ; lnx ..., , x (ln f
(r) ) 1 (
-
-
-
-
=
=
-
b
a x ln
exp
b
a x ln
exp b
)! r n (
! n ) i ( ) i (
r
1 i
1
.
b
a x ln
exp ) r n ( exp
) r (
-
- - (33)
Let , a
)
b
)
be the maximum likelihood estimators of a,
b, based on X
(1)
, X
(r),
and let V
1
= , b / ) a a ( -
)
V
2
= , b / b
)
and
W
i
= b / ) a X (ln
) i (
)
)
- (i=1, , r). It is easily shown that the
distributions of V
1
, V
2
are parameter-free, and that any r-
2 of the
W
i
s, say
W
1
,
W
r-2
, form a set of r-2
functionally independent ancillary statistics. We then find
in a straightforward manner that the joint density of V1,
V
2
, conditional on fixed
W=(
W
1
,
W
r-2
), is
) w ..., , w ; v , v ( f
2 - r 1 2 1
,
) v w v exp( ) r n (
) v w v exp( w v rv
exp v (
2 i 1
r
1 i
2 i 1
r
1 i
i 2 1
+ - -
+ - +
J =
= =
2 - r
2
) w (34)
where ) w
w
1
,
w
r
in
(34);
w
r-1
and
w
r
can be expressed as function of
w
1
,
w
r-2
only.
Let Y
(1)
be the smallest observation from an
independent second sample of km observations also from
the distribution (24). Writing V=(lnY
(1)
-a)/b and noting
that exp(V) is the smallest observation in a sample of size
km from the standard exponential distribution, we have
the density of V as
). v exp( km exp( ) v exp( km ) v ( f - = (35)
Since V is distributed independently of V
1
, V
2
we
find the joint density of V, V
1
, V
2
, conditional on
W=
w,
as the product of (34) and (35). Note that
b / ) a Y (ln W
) 1 (
)
)
- = =(V-V
1
)/V
2
; making the transformation
W=(V-V
1
)/V
2
, V
1
=V
1
, V
2
=V
2
, we find the joint density of
W, V
1
, V
2
, conditional on
W=
w, as
w)
; v , v , w ( f
2 1
+ + + J =
=
2
r
1 i
i 1
v w w v ) 1 r ( exp ) ( km w
)] wv v exp( km exp[
2 1
+ -
. ) v w exp( ) r n ( ) v w exp( ) v exp( exp
r
1 i
2 i 2 i 1
- + -
=
(36)
Now v
1
can be integrated out of (36) in a
straightforward way to give
w)
; v , w ( f
2
.
) v w exp( ) r n ( ) v w exp( ) wv exp( km
v w w exp v ) ( km
1 r
r
1 i
2 r 2 i 2
2
r
1 i
i
1 r
2
+
=
=
-
- + +
+ J
=
w
(37)
Consider, for fixed w (-<w<),
. dwdv ) ; v , w ( f } ; w W Pr{
2
w
2
0
= > w w (38)
A straightforward integration then gives us
>
s
d = >
w w ; w
Y
ln Pr } ; w W Pr{
) 1 (
)
)
( )
( ) ( )
- + +
=
-
s d
=
s d
s d
-
=
ds e ) r n ( e kme
e s
r
/ x ln s
r
1 i
/ x ln s
sw
0
/ x ln s
2 r
) r ( ) i (
r
1 i
) i (
)
)
)
)
)
)
( )
( ) ( )
. ds e ) r n ( e e s
1
0
r
/ x ln s
r
1 i
/ x ln s
/ x ln s
2 r ) r ( ) i (
r
1 i
) i (
-
-
s d
=
s d
s d
-
- +
=
)
)
)
)
)
)
(39)
The above expression holds for 3rn, with r=n
corresponding to complete (uncensored) sampling. In the
case r=2, , a
)
b
)
are jointly sufficient for a, b, so that it can
be considered the unconditional probability Pr(W>w). It
will be noted that in this case, the correct expression is
also given by (39), with r=2. Now the probability
statement
D
o
w
n
l
o
a
d
e
d
b
y
[
R
y
e
r
s
o
n
U
n
i
v
e
r
s
i
t
y
]
a
t
1
0
:
4
6
0
5
J
u
n
e
2
0
1
3
N.A. Nechval, K.N. Nechval, E.K. Vasermanis / AVIATION 2004, Vol VIII, No 3, 39
- 9 -
g =
>
s
d
w ; w
Y
ln Pr
) 1 (
)
)
(40)
leads to the warranty period (0, )) / w exp( d s
)
)
with
confidence level g, i.e. a lower 100g% prediction limit, L
g
,
for Y
(1)
is equal to )). / w exp( d s
)
)
For instance, consider the data of fatigue tests on a
particular type of structural component of the aircraft IL-
86. The data are for a complete sample of size r = n = 5,
with observations
Table. The Data of Fatigue Tests on a Particular Type of
Structural Component of IL-861 Aircraft
Observations
Time to crack initiation
(in number of 10
4
flight-hours)
) 1 (
x 5
) 2 (
x
25 . 6
) 3 (
x
5 . 7
) 4 (
x
9 . 7
) 5 (
x
1 . 8
The results are being expressed here in number of
10
4
flight-hours. On the basis of these data, the wish is to
estimate a lower 0.95 prediction limit on Y
(1)
in a group of
m = 5 identical components (for a fleet of k=1 IL-86
aircraft) that are to be put into service.
Goodness-of-fit testing. We assess the statistical
significance of departures from the Weibull model by
performing the empirical distribution function goodness-
of-fit test. We use the S statistic [3]. For censoring (or
complete) data sets, the S statistic is given by
, 184 . 0
M
) x / x ln(
M
) x / x ln(
M
) x / x ln(
M
) x / x ln(
S
4
1 i i
) i ( ) 1 i (
4
3 i i
) i ( ) 1 i (
1 r
1 i i
) i ( ) 1 i (
1 r
1 ] 2 / r [ i i
) i ( ) 1 i (
=
=
+
=
+
-
=
+
-
+ =
+
(41)
where [r/2] is a largest integer r/2, the values of M
i
are given in Table 13 [3]. The reject region for the a level
of significance is {S>S
n;1-a
}. The percentage points for S
n;1-
a
were given by Kapur and Lamberson [3]. For this
example,
S=0.184 < S
n=5; 1-a=0.95
=0.86. (42)
Thus, there is no evidence to rule out the Weibull
model.
The maximum likelihood estimates are
42603 . 7 = s
)
and . 9081 . 7 = d
)
It follows from (39) that
95 . 0
0000170442 . 0
000016192 . 0
55761 . 4
Y
ln Pr
) 1 (
= =
- >
s
d
)
)
(43)
and a lower 0.95 prediction limit for Y
(1)
is 4.1730
(10
4
) flight-hours, i.e. we have obtained the warranty
period equal to 41730 flight-hours with confidence level
g=0.95.
Conclusions
In this paper we consider the important situation in
which it can be assumed that the structural components of
the aircraft in question have the time to crack initiation
following the Weibull distribution. It will be noted that
the general problem considered here, that of predicting on
the basis of an ordered sample the smallest observation
Y
(1,km)
from k future samples, each consisting of m units,
has application in reliability theory other than described
above. For example, if one has a series system consisting
of m identical components, with lifetimes Y
1
, Y
m
, then
Y
(1)
Y
(1,m)
represents the life of the system; it is often
wished to estimate Y
(1)
for a given system, on the basis of
previous life test data on the components.
Acknowledgments
The authors wish to acknowledge the partial support
of this research by the Latvian Council of Science and the
National Institute of Mathematics and Informatics of
Latvia under Grant No. 02.0918 and Grant No. 01.0031.
References
1. Chand S., Garg S.B.L. Propagation Under Constant
Amplitude Loading // Engineering Fracture
Mechanics. 1985. Vol. 21. P. 130.
2. Hoeppner D.W., Krupp W.E. Prediction of
Component Life by Application of Fatigue Crack
Growth Knowledge // Engineering Fracture
Mechanics. 1974. Vol. 6. P. 4770.
3. Kapur K.C., Lamberson L.R. Reliability in
Engineering Design. John Wiley & Sons, 1977.
4. Nechval N.A. Modern Statistical Methods of
Operations Research. Riga: RCAEI, 1982.
5. Nechval N.A. Theory and Methods of Adaptive
Control of Stochastic Processes. Riga: RCAEI,
1984.
6. Nechval N.A., Nechval K.N. State Estimation of
Stochastic Systems via Invariant Embedding
7. Technique // Cybernetics and Systems2000 / Edited
by R. Trappl. Vienna: Austrian Society for
Cybernetic Studies, 2000. Vol. 1. P. 96101.
8. Nechval N.A., Nechval K.N., Vasermanis E.K.
Invariant Embedding Technique and its Applications
to Statistical Decision-Making // Proceedings of the
Second World Congress of Latvian Scientists. Riga:
LZP, 2001. P. 578.
9. Nechval N.A., Nechval K.N., Vasermanis E.K.
Optimization of Interval Estimators via Invariant
Embedding Technique // IJCAS (International Journal
of Computing Anticipatory Systems). 2001. Vol 9.
P. 241255.
10. Schwalbe K-H. Comparison of Several Fatigue Crack
Propagation Laws With Experimental Results //
Engineering Fracture Mechanics. 1974. Vol 6. P.
325341.
D
o
w
n
l
o
a
d
e
d
b
y
[
R
y
e
r
s
o
n
U
n
i
v
e
r
s
i
t
y
]
a
t
1
0
:
4
6
0
5
J
u
n
e
2
0
1
3