r
i=1
X
i
as
the total number of arrivals before time r and T
nr
=
n
i=r+1
X
i
as the total number
of arrivals after time r. T
0
0 and T
0
= 0. Thus, when t = r, T
r
is a Poisson (rz
1
)
D
o
w
n
l
o
a
d
e
d
A
t
:
0
9
:
4
7
2
5
J
u
n
e
2
0
1
1
70 Brown
and T
nr
is Poisson ((n r)z
2
). If (z
1
. z
2
) are both known, the likelihood of t is
given by
L (t
n
) =
n1
r=0
I (t = r)z
T
r
1
z
T
nr
2
e
z
1r
z
2
(nr)
+I (t n)e
z
1
n
z
T
n
1
. (2.1)
When (z
1
. z
2
) are unknown, the prole likelihood function of t is the expected
value of (2.1) with respect to the prior distribution of (z
1
. z
2
). We assume a priori
that (z
1
. z
2
) is uniformly distributed in some interval 0 - z
1
- z
2
- M. That is, we
know that when the change happens, the arrival rates have increased, and we also
know that the arrival rates are bounded and uniform in a certain interval. Thus the
prole likelihood function of t, given
n
, is
L
(t
n
) = I (t = 0)
M
0
z
2
0
e
z
2
n
z
T
n
2
Jz
1
Jz
2
+
n1
r=1
I (t = r)
M
0
z
2
0
z
T
r
1
z
T
nr
2
e
z
1
rz
2
(nr)
Jz
1
Jz
2
+I (t n)
M
0
z
2
0
e
z
1
n
z
T
n
1
Jz
1
Jz
2
. (2.2)
For r = 0, let
L
0
=
M
0
z
2
0
e
z
2
n
z
T
n
2
Jz
1
Jz
2
=
(T
n
+1)!
n
T
n
+2
P(T
n
+1; Mn). (2.3)
where P(X; z) is the survival function of the Poisson distribution with parameter z.
For 1 r n 1, let
L
r
=
M
0
z
T
nr
2
e
z
2
(nr)
z
2
0
z
T
r
1
e
z
1
r
Jz
1
Jz
2
=
T
r
!
r
T
r
+1
M
0
z
T
nr
2
e
z
2
(nr)
P(T
r
; rz
2
)Jz
2
= T
r
!
;=0
r
;
(T
n
+; +1)!
(T
r
+; +1)!n
T
n
+; +2
P(T
n
+; +1; Mn). (2.4)
Finally, let
L
n
=
M
0
M
z
1
e
z
1
n
z
T
n
1
Jz
2
Jz
1
=
T
n
!
n
T
n
+1
MP(T
n
; Mn)
T
n
+1
n
P(T
n
+1; Mn)
. (2.5)
We put Shiryaevs geometric prior distribution h on the change point t,
h (t = i) =
for i = 0
(1 )(1 )
i1
for i 1.
(2.6)
D
o
w
n
l
o
a
d
e
d
A
t
:
0
9
:
4
7
2
5
J
u
n
e
2
0
1
1
Bayesian Detection of Changes of a Poission Process 71
Thus, we obtain that the posterior probability that there has been a change by time
n is
t
n
=
L
0
+(1 )
n1
r=0
(1 )
r
L
r
L
0
+(1 )
n1
r=0
(1 )
r
L
r
+(1 )(1 )
n
L
n
. (2.7)
Let
t
n
= 1 t
n
be the probability that there has not been a change by time n.
3. OPTIMAL STOPPING RULE
In this section, we nd an optimal stopping rule. Suppose the cost of stopping early
is c
1
loss units and the cost per time unit of stopping late is c
2
loss units. We want
to nd a stopping rule that will minimize the expected cost. We assume without loss
of generality that c
1
= 1 and c
2
= c is the relative cost of stopping late. Let R
n
be
the risk (or minimal expected loss) at time n. Therefore,
R
n
(
n
) = min(
t
n
. ct
n
+E(R
n+1
n
)). (3.1)
The stopping rule, then, would be to stop sampling when the risk of stopping
is smaller than the risk of continuing. According to (3.1), this rule is equivalent to
saying we stop sampling at the rst n such that
t
n
- ct
n
+E(R
n+1
n
). (3.2)
In order to calculate R
n
, we use the principle of dynamic programming. Suppose
we must stop after n
= n.
Therefore at time n we must stop and our risk is
R
(0)
n
(
n
) =
t
n
. (3.3)
In general, for ; 1, we have a choice: we can stop and say that a change
has occurred, or we can continue and take another observation. If we decide to
continue, our risk is
R
(;)
n
= ct
n
+E
R
(;1)
n+1
n
t
n
.
R
(;)
n
). We can further write this as
R
(;)
n
=
t
n
+
R
(;)
n
t
n
=
t
n
+
c
t
n
(c +1) +E
R
(;1)
n+1
n+1
.
(3.4)
Theorem 3.1. The sequence ](
t
n
.
n
). n 1] is a super martingale.
Proof. Taking the conditional expectation of t
n+1
given the o-eld generated by
the rst n observations, we obtain
E(t
n+1
n+1
) = E(E(I (t - n +1)
n+1
)
n
)
= E(E(I (t - n)
n+1
)
n
) +E(E(I (t = 1)
n+1
)
n
)
=
t
n
+
t
n
. (3.5)
D
o
w
n
l
o
a
d
e
d
A
t
:
0
9
:
4
7
2
5
J
u
n
e
2
0
1
1
72 Brown
Therefore,
E
t
n+1
n+1
= (1 )
t
n
-
t
n
a.s. (3.6)
Accordingly, the sequence ]
t
n
. n 0] is a super martingale.
Corollary 3.1.
lim
n
t
n
= 0 a.s (3.7)
Proof. From equation (3.6),
E
t
n+1
n+1
= (1 )
t
n
= (1 )
n
(1 ). (3.8)
Thus, the lim sup
n
E(
t
n
) = 0. Since
t
n
0 with probability 1 for all n, we
obtain (3.7).
Thus, the one-step look-ahead procedure states that we stop sampling and
declare that a change has happened at the rst n such that
t
n
- q
=
c
c +
. (3.9)
Dene M
(;)
n
= E
R
(;1)
n+1
t
n+1
for ; 1, and M
(0)
n
0. Thus, the functions M
(;)
n
satisfy the recursive relationship
M
(;)
n
= E
c
t
n+1
(c +) +M
(;1)
n+1
. (3.10)
Hence, the risk can be written as a function of M
(;)
n
, namely,
R
(;)
n
=
t
n
+|c
t
n
(c +) +M
(;1)
n
]
. (3.11)
We then make the decision to stop when c
t
n
(c +) +M
(;1)
n
> 0, which is
equivalent to stopping when
t
n
- q
+
M
(;1)
n
c+
. We can dene the ;th-step-ahead
boundary to be l
(;)
n
= max
0. q
+
M
(;1)
n
c+
.
Lemma 3.1. lim
;
l
(;)
n
= l
n
exists. That is, there exists an optimal stopping
boundary.
Proof. By induction one can show that M
(;)
n
are decreasing in ;. Thus, l
(;)
n
are
decreasing functions of ;. Also, l
(;)
n
> 0 for all ;. By the monotone convergence
theorem, the limit exists.
Lemma 3.2. lim
n
l
(;)
n
= q
.
Proof. Let
I =
i c
t
n+1
(X
n
. i)(c +) > 0
. (3.12)
Therefore, if i I. then
t
n+1
- q
t
n+1
(X
n
. i) - q
] = 1. (3.13)
Therefore, lim
n
M
(1)
n
= 0. Thus, lim
n
l
(2)
n
= q
c
t
n
(c +) +M
(1)
n
. (4.1)
To evaluate the two-step-ahead risk, we nd an expression for M
(1)
n
. First, we express
t
n+1
recursively as a function of
t
n
:
t
n+1
t
n
(1 )
V
n
D
n
(
n
)
D
n+1
(
n+1
)
. (4.2)
where, according to (2.5),
V
n
=
L
(n+1)
n+1
L
(n)
n
(4.3)
and L
(n)
r
is dened as L
r
when the likelihood is being calculated with respect to
n
,
the o-eld generated by the rst n observations. According to (3.10),
M
(1)
n
= E
c
t
n+1
(c +)
. (4.4)
We only need to take the expectation over those X
n+1
for which
t
n+1
> q
.
Lemma 4.1. If X
n+1
M, then
t
n+1
-
t
n
.
Proof. Let X
n+1
= m M. Let W
r
=
L
(n+1)
r
L
(n)
r
. For r = 0,
W
0
=
(T
n
+m+1)!n
T
n
+2
P(T
n
+m+1; M(n +1))
(T
n
+1)!(n +1)
T
n
+m+2
P(T
n
+1; Mn)
. (4.5)
If we divide equation (4.5) by
V
n
given in (4.3), we obtain,
W
0
V
n
=
E
G
, where E and
G are dened as
E = P(T
n
+m+1; M(n +1))
Mn
T
n
+1
P(T
n
; Mn) P(T
n
+1; Mn)
G = P(T
n
+1; Mn)
M(n +1)
T
n
+m+1
P(T
n
+m; M(n +1)) (4.6)
P(T
n
+m+1; M(n +1))
We show that W
0
>
V
n
. This is equivalent to showing that the rst term in E is
greater than the rst term in G. (See Lemma A.1 Appendix for details). Therefore,
W
0
V
n
> 1. Now, let 0 - r - n:
W
r
=
;
r
;
(T
n
+;+m+1)!P(T
n
+m+;+1;M(n+1))
(T
r
+;+1)!(n+1)
T
n
+;+m+1
;
r
;
(T
n
+;+1)!P(T
n
+;+1;Mn)
(T
r
+;+1)!n
T
n
+;+1
. (4.7)
D
o
w
n
l
o
a
d
e
d
A
t
:
0
9
:
4
7
2
5
J
u
n
e
2
0
1
1
74 Brown
To show that W
r
>
V
n
, it sufces to show that each term in the numerator times
V
n
is greater than the corresponding term in the denominator of (4.7) for each ;. r,
and n. This is equivalent to showing E
;
> G
;
, where E
;
and G
;
axe given by
E
;
=
MT
n
!(T
n
+; +m+1)!
(n +1)
;+1
P(T
n
+; +m+1; M(n +1))P(T
n
; Mn)
(T
n
+1)!(T
n
+; +m+1)!
n(n +1)
;+1
P(T
n
+; +m+1; M(n +1))P(T
n
+1; Mn); (4.8)
G
;
=
(T
n
+; +1)!(T
n
+m)
n
;+1
P(T
n
+; +1; Mn)P(T
n
+m; M(n +1))
(T
n
+; +1)!(T
n
+m+1)!
(n +1)n
;+1
P(T
n
+; +1; Mn)P(T
n
+m+1; M(n +1)). (4.9)
The inequality E
;
> G
;
can be shown by induction on ;, making use of
Lemma A.1 in the appendix. Therefore, W
r
>
V
n
for all r - n. Thus,
t
n+1
t
n
=
V
n
(1 )
L
(n)
0
+(1 )
n1
r=0
(1 )
r
L
(n)
r
+(1 )(1 )
n
L
(n)
n
L
(n+1)
0
+(1 )
n
r=0
(1 )
r
L
(n+1)
r
+(1 )(1 )
n+1
L
(n+1)
n+1
.
(4.10)
From (4.8) and (4.9), we see that L
(n+1)
r
= W
r
L
(n)
r
>
V
n
L
(n)
r
. Therefore,
t
n+1
t
n
-
(1 )
V
n
D
n
(
n
)
V
n
D
n
(
n
) +(1 )(1 )
n
L
(n+1)
n
+L
(n+1)
n+1
- 1. (4.11)
Therefore, if the (n +1)th observation is greater than M, the posterior probability
that a change has not occurred will decrease.
We use Lemma 4.1, to evaluate the two-step-ahead stopping rule. According
to equation (4.1), the two-step-ahead risk function is R
(2)
n
=
t
n
+
c
t
n
(c +) +
M
(1)
n
where M
(1)
n
is dened as M
(1)
n
= E||c
t
n+1
(c +)]
. Note that if
t
n
- q
, then
t(X
n
. k) - q
,
M
(1)
n
=
i-M
D
n+1
(X
n
. i)
D
n
(X
n
)
c (c +)
t
n
(1 )
V
n
(i)D
n
(X
n
)
D
n+1
(X
n
. i)
=
i-M
c
D
n+1
(X
n
. i)
D
n
(X
n
)
(c +)
t
n
(1 )
V
n
(i)
(4.12)
Thus the two-step-ahead boundary is
B
n.2
=
+
i-M
D
n+1
(X
n
. i)
D
n
(X
n
)
t
n
(1 )
V
n
+
(4.13)
Notice B
n.2
= l
(2)
n
when
t
n
- q
. Also when
t
n
> q
for
large n. This illustrates Lemma 3.2.
From looking at Table 1 one observes that we would stop at time 11. Since the
change took place at time 10, this shows that the detection rule is quick. We also
see the convergence between the one- and two-step-ahead stopping rules.
APPENDIX A
Lemma A.1. For all n. k, and m positive integers, and m M > 0,
k +m
n +1
P(k +m; M(n +1))P(k 1; Mn) >
k
n
P(k +m1; M(n +1))P(k; Mn). (A.1)
Proof. Taking differences and utilizing the relationship between the cumulative
distribution function of the Poisson distribution and the gamma distribution,
D
o
w
n
l
o
a
d
e
d
A
t
:
0
9
:
4
7
2
5
J
u
n
e
2
0
1
1
76 Brown
we obtain
k +m
n +1
P(k +m; M(n +1))P(k 1; Mn)
k
n
P(k +m1; M(n +1))P(k; Mn)
=
k +m
n +1
1
0
M
k+m+1
(n +1)
k+m+1
(k +m)!
x
k+m
e
M(n+1)x
1
0
M
k
n
k
(k 1)!
x
k1
e
Mnx
Jx Jx
k
n
1
0
M
k+m
(n +1)
k+m
(k +m1)!
x
k+m1
e
M(n+1)x
1
0
M
k+1
n
k+1
k!
x
k
e
Mnx
Jx Jx. (A.2)
Simplifying, we obtain that equation (A.2) is equal to
M
2k+m+1
1
0
(n +1)
k+m
(k +m1)!
x
k+m1
e
M(n+1)x
1
0
n
k
(k 1)!
x
k1
e
Mnx
(x x)Jx Jx. (A.3)
We will divide the integral into where x - x and a second integral where x > x.
Thus, equation (A.3) is equivalent to M
2k+m+1
times
1
0
(n +1)
k+m
(k +m1)!
x
k+m1
e
M(n+1)x
x
0
n
k
(k 1)!
x
k1
e
Mnx
(x x)Jx Jx
+
1
0
(n +1)
k+m
(k +m1)!
x
k+m1
e
M(n+1)x
1
x
n
k
(k 1)!
x
k1
e
Mnx
(x x)Jx Jx. (A.4)
Note that the second integral in (A.4) can be rewritten as
1
0
(n +1)
k+m
(k +m1)!
x
k+m1
e
M(n+1)x
1
x
n
k
(k 1)!
x
k1
e
Mnx
(x x)Jx Jx
=
1
0
x
0
M(n +1)
k+m
(k +m1)!
x
k+m1
e
M(n+1)x
n
k
(k 1)!
x
k1
e
Mnx
(x x)Jx Jx. (A.5)
Thus, equation (A.2) is
k +m
n +1
P(k +m; M(n +1))P(k 1; Mn)
k +m
n
P(k; M(n +1))P(k +m; Mn)
= M
2k+m+1
1
0
x
0
(n +1)
k+m
n
k
k!(k 1)!
x
k1
x
k1
e
Mn(x+x)
(x
m
e
Mx
x
m
e
Mx
)Jx Jx. (A.6)
g(x) = x
m
e
Mx
is an increasing function over (0,1) when m > M. Hence,
(n +1)
k+1
n
k
k!(k 1)!
x
k1
x
k1
e
n(x+x)
(x
m
e
Mx
x
m
e
Mx
) > 0 (A.7)
for all 0 - x - x - 1. Therefore,
k +m
n +1
P(k +m. M(n +1))P(k 1; Mn) >
k
n
P(k +m1; M(n +1))P(k; Mn)
(A.8)
for all n. k. m, and M such that m M.
D
o
w
n
l
o
a
d
e
d
A
t
:
0
9
:
4
7
2
5
J
u
n
e
2
0
1
1
Bayesian Detection of Changes of a Poission Process 77
ACKNOWLEDGMENTS
I would like to thank Dr. Shelemyahu Zacks for his helpful advice in preparing this
article. I would also like to thank the Editor, Associate Editor, and referees for their
helpful suggestions.
REFERENCES
Brown, M. and Zacks, S. (2006a). Note on Optimal Stopping for Possible Change
in the Intensity of an Ordinary Polsson Process, Probability and Statistics Letters
76: 14171425.
Brown, M. and Zacks, S. (2006b). On the Bayesian Detection of a Change in the Arrival
Rate of a Poisson Process Monitored at Discrete Epochs, International Journal for
Statistics and Systems 1: 113.
Chernoff, H. and Zacks, S. (1964). Estimating the Current Mean of a Normal Distribution
Which is Subjected to Changes in Time, Annals of Mathematical Statistics 35: 9991018.
Herberts, T. and Jensen, U. (2004). Optimal Detection of a Change Point in a
Poisson Process for Different Observation Schemes, Scandinavian Journal of Statistics
31: 347366.
Hinkley, D. and Hinkley, E. (1970). Inference about the Change-Point in a Sequence of
Binomial Variables, Biometrika 57: 477488.
Kander, Z. and Zacks, S. (1966). Test Procedures for Possible Changes in Parameters of
Statistical Distributions Occurring at Unknown Time Points, Annals of Mathematical
Statistics 37: 11961210.
Page, E. S. (1954). Continuous Inspection Schemes, Biometrika 41: 100114.
Page, E. S. (1962). Cumulative Sum Schemes Using Gouging, Technometrics 4: 97109.
Peskir, G. and Shiryaev, A. N. (2002). Solving the Poisson Disorder Problem, in Advances
in Finance and Stochastics: Essays in Honor of Dieter Sonderman, K. Sandmann and
P. J. Schonbucher, eds., pp. 295312, New York: Springer.
Shewhart, W. A. (1926). Quality Control Charts, Bell System Technical Journal 5: 593603.
Shiryaev, A. N. (1978). Statistical Sequential Analysis: Optimal Stopping Rules, in
Translations of American Mathematical Society 38, Providence: American Mathematical
Society, 149.
Smith, A. F. M. (1975). A Bayesian Approach to Inference about a Change-Point in the
Sequence of Random Variables, Biometrika 62: 407416.
Zacks, S. (1983). Survey of Classical and Bayesian Approaches to the Change-Point
Problem: Fixed Sample and Sequential Procedures for Testing and Estimation,
in Recent Advances in Statistics, M H. Rizvi, J. Rastogi, and D. Siegmund, eds.,
pp. 245269, New York: Academic Press.
Zacks, S. and Barzily, Z. (1981). Bayes Procedures for Detecting a Shift in the Probability
of Success in a Series of Bernoulli Trials, Journal of Statistical Planning and Inference
5: 107119.
D
o
w
n
l
o
a
d
e
d
A
t
:
0
9
:
4
7
2
5
J
u
n
e
2
0
1
1