I.
INTRODUCTION
2
which has been widely used to characterize the entangled
states since Aspects experiment [1], measures the entanglement by the ratios obtained at two fixed bases for one
particle. We will show that the visibility is not sufficient
and to be re-formulated in the view of statistics. We then
improve the test by optimizing the allocation on the measurement time for each measurement basis, considering
that the counts on the anti-coincidence bases are much
smaller than those on the coincidence bases. The test
can be further improved, if we utilize the knowledge on
the tendency of the entanglement degradation. In general, the error from the maximally entangled states can
be anisotropic, which reflects the generation process of
the states. We can improve the sensitivity to the entanglement degradation by focusing the measurement on the
expected error directions.
The SPDC is now widely used to generate entangled
photon pairs. Several experimental settings have been
demonstrated to provide highly entangled states. In particular, Kwait et al [15] have obtained a high flux of the
polarization entangled photon pairs from a stack of two
type-I phase matched nonlinear crystals. One nonlinear
crystal generates a photon pair polarized in the horizontal direction (|HHi), and the other generates ones polarized in the vertical direction (|V V i). If the two pairs are
indistinguishable, the generated photons are entangled.
Otherwise, the state will be a mixture of HH pairs and
V V pairs. The quantum state tomography has shown
that the only HHHH, V V V V , V V HH, HHV V elements are dominant [16], which implies that the density
matrix can be approximately given by the classical mixture of the |(+) ih(+) | and |() ih() |. We can improve the entanglement test based on this property of
the photon pairs, as described in the following sections.
In this article, we reformulate the hypothesis testing
to be applicable to the SPDC experiments, and demonstrate the effectiveness of the optimized time allocation
in the entanglement test. The construction of this article is following. Section 2 defines the hypothesis scheme
for the entanglement of the two-photon states generated
by SPDC. Section 3 gives the mathematical formulation
concerning statistical hypothesis testing. Sections 4 - 10
describe the mathematical aspects, and Section 11 examines the experimental aspects of the hypothesis testing on
the entanglement. The designs on the time allocation are
evaluated by the experimental data. Hence, if a reader is
interested only in experiments and data analysis, he can
skip sections 4 - 10 and proceed to section 11. However,
if he is concerned with the mathematical discussion, he
should read sections 4 - 10 before section 11.
The following is organization of sections 4 - 10, which
discuss more theoretical issues. Sections 4 and 5 give the
fundamental properties of the hypothesis testing: section
4 introduces the likelihood ratio test, and section 5 gives
the asymptotic theory of the hypothesis testing. Sections
6-9 are devoted to the designs of the time allocation between the coincidence and anti-coincidence bases: section
6 defines the modified visibility method, section 7 opti-
II.
This section introduces the hypothesis test for entanglement. We consider the two-photon states generated by
SPDC, which are described by a density matrix . We
assume each two-photon generation process to be identical but individual. The target state is the maximally
entangled |(+) i state. Here we measure the entanglement by the fidelity between the generated state and the
target state:
= h(+) ||(+) i.
(1)
(2)
with a threshold 0 .
The outcome of the coincidence count measurement in
the SPDC experiment can be assumed a random variable independently distributed according to the Poisson
distribution. From now on, a symbol labeled by a pair
(x, y) refers to a random variable or parameter related to
the measurement basis |xA , yB i. The number of detection events nxy on the basis |xA , yB i is a random variable
according to the Poisson distribution Poi((xy + )txy )
of mean (xy + )txy , where
is a known constant, related to the photon detection rate, determined from the average number of
photon pairs generated in unit time and the detection efficiency,
xy = hxA , yB ||xA , yB i is an unknown constant,
txy is a known constant of the time for detection.
is a constant of the average dark counts,
The probability function of nxy is
exp((xy + )txy )
3
Because the detections at different times are mutually
independent, nxy is independent of nx y (x 6= x or y 6=
y ). In this paper, we discuss the quantum hypothesis
testing under the above assumption while Usami et al.[17]
discussed the state estimation under this assumption.
Visibility of the two photon interference is an indicator of entanglement commonly used in the experiments.
When the dark-count parameter can be regarded to
be 0, the visibility is calculated as follows: first, As
measurement basis |xA i is fixed, then the measurement
|xA , yB i is performed by rotating Bs measurement basis |xB i to obtain the maximum and minimum number
of the coincidence counts, nmax and nmin . We need to
make the measurement with at least two bases of A in
order to exclude the possibility of the classical correlation. We may choose the two bases |Hi and |Di as |xA i,
for example. Finally, the visibility is given by the ratio
between nmax nmin and nmax + nmin with the respective As measurement basis |xA i. However, our decision
will contain a bias, if we choose only two bases as As
measurement basis |xA i. Hence, we cannot estimate the
fidelity between the target maximally entangled state and
the given state in a statistically proper way from the visibility.
Since the equation
|HHihHH| + |V V ihV V | + |DDihDD|
+ |XXihXX| + |RLihRL| + |LRihLR|
=2|(+) ih(+) | + I
(3)
(4)
Formulation
(5)
4
probability, we maximize the probability to reject the hypothesis H0 when the true parameter is 1 . This
probability is given as P (), and is called the power of
. Hence, a test of the risk probability is said to be
most powerful (MP) at 1 if P () P () holds
for any test of the risk probability . Then, a test is
said to be Uniformly Most Powerful (UMP) if it is MP
at any 1 .
B.
p-values
In the hypothesis testing, we usually fixed our test before applying it to data. However, we sometimes focus
on the minimum risk probability among tests in a class T
rejecting the hypothesis H0 with a given data. This value
is called the p-value, which depends on the observed data
x as well as the subset 0 to be rejected.
In fact, in order to define the p-value, we have to fix a
class T of tests. Then, for x and 0 , p-value is defined
as
min
max p ().
T :(x)=1 0
Definition
(8)
+ rP1 () = P0 () + rP1 ()
P0 (r ) + rP1 (r ) = + rP1 (r ).
Hence, 1 P1 () 1 P1 (r ). This is known as
Neyman-Pearsons fundamental lemma[19].
The likelihood ratio test is generalized to the cases
where 0 or 1 has at least two elements as
0 if sup0 p (x) r,
sup1 p (x)
r (x) :=
1 if sup0 p (x) < r.
p (x)
sup
1
B.
(9)
(6)
Since the p-value expresses the risk for rejecting the hypothesis H0 , Hence, this concept is useful for comparison
among several experiments.
Note that if we are allowed to choose any function
as a test, the above minimum is attained by the function
x :
0 if y 6= x
x (y) =
(7)
1 if y = x.
IV.
(10)
we often use so-called interval tests for its optimality under some conditions as well as for its naturalness.
When the likelihood ratio P (x)/P (x) is monotone
increasing concerning x for any , such that > ,
the likelihood ratio is called monotone. In this case, the
likelihood test r between P0 and P1 is UMP of level
:= P0 (r ), where 1 is an arbitrary element satisfying
1 < 0 .
Indeed, many important examples satisfy this condition. Hence, it is convenient to give its proof here.
From the monotonicity, the likelihood test r has the
form
1 x < x0
(11)
r (x) =
0 x x0
with a threshold value x0 . Since the monotonicity implies
P0 (r ) P (r ) for any 0 , it follows from Neyman
Pearson Lemma that the likelihood test r is MP of level
. From (11), the likelihood test r is also a likelihood
test between P0 and P , where is another element
satisfying < 0 . Hence, the test r is also MP of level
.
From the above discussion, it is suitable to treat pvalue based on the class of likelihood tests. In this case,
when we observe x0 , the p-value is equal to
Z x0
(12)
P0 (dx).
C.
5
exponential family when there exists a random variable
x such that
P (x) := P0 (x) exp(x + g()),
(13)
R
where g() := log exp(x)P0 (dx).
It is known that this class of families includes, for example, the Poisson distributions, normal distributions,
binomial distributions, etc. In this case, the likelihood
0 x+g(0 ))
ratio exp(
exp(1 x+g(1 )) = exp((0 1 )x + g(0 ) g(1 )) is
monotone concerning x for 0 > 1 . Hence, the likelihood
ratio test is UMP in the hypothesis (10). Note that this
argument is valid even if we choose a different parameter
if the family has a parameter satisfying (13).
For example, in the case of the normal distribution
(x)2
x2
1
1
e 2V = 2V
e 2V + V 2V , the UMP
P (x) = 2V
test UMP, of the level is given as
(
1 if x < 0 V
(14)
UMP, (x) :=
0 if x 0 V ,
1
1
0 versus H1 :
< 0 .
1 + 2
1 + 2
x2
1
e 2 dx.
2
(15)
The n-trial binomial distributions Ppn (k) = nk (1
p)nk pk are also an exponential family because anp
satisfies that Ppn (k) =
other parameter := log 1p
e
n 1 k+n log 1+e
. Hence, in the case of the n-trial bik 2n e
nomial distribution, the UMP test nUMP, of the level
is given as
1 if x < k0
UMP, (k) := if x = k0
(16)
0 if x > k
0
(17)
(20)
k1 +k2
In this case, the test (k1 , k2 ) = UMP,
(k1 ) is a test
of level . This is because the conditional distribution
P n Poi(1 ,2 )(k,nk)
Poi(1 ,2 )(k ,nk ) is equal to the binomial distribuk =0
tion P n 1 (k). Therefore, when we observe k1 , k2 , the p1 +2
Pk1 k1 +k2 k
0 (1
value of this class of tests is equal to k=0
k
)k1 +k2 k .
D.
where
Multi-parameter case
(21)
(22)
6
This is because the logarithm of the likelihood function
is calculated as
sup0 0 P0 (x)
sup1 1 P1 (x)
P (x)
= sup inf log 0
P1 (x)
1
1
0 0
log
= inf
c<c0
0 0 1 1
= sup
0 0 1 1
= sup
0 0 1 1
1 1
(23)
1
(2)m/2 det V
e 2 xV
D(P kP ) =
1
2 (
x+V 1 x 21 V 1
)V
(wc)2
2wV 1 w ,
the likelihood
1 1
= sup
minw=c0 21 ( )V 1 ( ) =
function is calculated as
0 0
(w x c)
(w x c)2
inf
.
1
cc0 2wV 1 w
2wV w
That is, the likelihood function has the same form as the
likelihood function of one-parameter normal distribution
family with the variance wV 1 w.
The
multi-nomial
Poisson
distributions
k
Pl
m
1 k
l
Poi(1 , . . . , m )(k1 , . . . , km )
:=
e i=1 i k11 !km
!
is also an exponential family.
The divergence is
calculated as
D(Poi(1 , . . . , m )kPoi(1 , . . . , m ))
m
m
X
X
i
i log .
(i i ) +
=
i
i=1
i=1
)
Pm
Pm
ki
k
)
+
k
log
min
i
i=1 i
i
,
max Poi(1 , . . . , m ) (k1 , . . . , km
) Pmw=c0 i=1 i
w =c0
i=1 wi ki < c0
max
w =c0
Poi(1 , . . . , m )
k
)
X k
i
1 ,
i=1 i
(27)
(k1 , . . . , km
)
where
i is defined as follows:
V.
where
wM
c0 (wM wi )
wi wM
log
:=
maxi wi
wM wi
wM .
and
R0
:=
(28)
(29)
c0
wM
(26)
ASYMPTOTIC THEORY
A.
Fisher information
c0
i wi
i +
i log
= R if R R0
wi
c0
wM wi
c0
+
i log
= R if R > R0 ,
wM
wM
(25)
Pm
Pm
where R := minw=c0 i=1 (i ki )+ i=1 ki log kii . As
is shown in Appendix D, this value is upperly bounded
by
(24)
7
the Fisher information J0 at the threshold 0 .
In the case of Poisson distribution family Poi(t), the
parameter can be estimated by Xt . The asymptotic case
corresponds to the case with large t. In this case, Fisher
information is t . When X obey the unknown Poisson
distribution family Poi(t), the estimation error Xt is
close to the normal distribution with the variance t , i.e.,
X
t( t ) approaches to the random variables obeying
the normal distribution with variance . That is, Fisher
information corresponds to the inverse of variance of the
estimator.
This approximation can be extended to the multiparameter case {p | Rm }. Similarly, it is known that
the testing problem can be approximated by the testing
of the normal distribution family with the covariance matrix (nJ )1 , where the Fisher information matrix J;i,j
is given by
Z
J;i,j := l;i (x)l;j (x)P (dx),
(31)
l;i (x) :=
log p (x)
.
i
(x) :=
wJ 1 w
0
,
distribution family with variance
n
Indeed, the same fact holds for the multinomial Poisson
distribution family Poi(t1 , . . . , tm ). When the random
variable Xj is the i-th random variable, the random variPm j
able
j=1 t (Xj j ) converges to the random variable
the normal distribution with the variance
Pm obeying
2
j=1 j j in distribution:
m
m
X
X
j
d
(Xj j )
2j j .
t
j=1
j=1
(33)
This convergence is compact uniform concerning the parameter 1 , . . . , m . In this case, the Fisher information
matrix J is the diagonal matrix with the diagonal elements ( t1 , . . . , tm ). When our distribution family is
given as a subfamily Poi(t1 (), . . . , tm ()), the Fisher
(34)
x2
1
e 2 dx.
2
In the n-trial binomial distribution case, the distribution can approximated by normal distribution if n is sufficiently large. In this case, the p-value of likelihood ration
tests is the function of the data k, the number of trial and
the threshold 0 , which is equal to
k n0
( p
).
n(1 0 )0
(35)
(32)
When the hypotheses is given by (10), the testing problem can be approximated by the testing of the normal
B.
distributions with the variance v. In this case, the pvalue of likelihood ration tests is the function of the data
x, the variance v, and the threshold 0 , which is equal to
0 ), where
( x
v
n 0 t
),
(
0 t
(36)
(37)
Pm
Pm
1 j=1 jj
1 j=1 jj
= max q
,
max qP
Pm i
m i
w =c0
w =c0
i=1
2i
i=1
2i
(38)
because this convergence (33) is compact uniform concerning the parameter 1 , . . . , m . Letting xi = wci0i
and yi = wci 02 , we have
i
Pm
1 j=1 jj
1x
max qP
= max ,
m i
w =c0
y
(x,y)Co
i=1
(39)
2i
(x,y)Co
1x
).
y
(40)
8
VI.
MODIFICATION OF VISIBILITY
(2 + 1)(2 2)
.
=
t
t
3(22) )
t1 t2
3
(42)
2
,
1 + 2
22
where 1 = 2+1
12 t and 2 = 12 t are the expectation values of the total coincidence counts and the total anti-coincidence counts, respectively. Considering the
definition of visibility, (nmax nmin )/(nmax + nmin ) =
(2nmin /(nmax + nmin )), we can regard the above estimation of fidelity as a modification of the visibility in a
well-defined statistics manner. We will refer to it as the
modified visibility method. In the following, we will propose several designs of experiment to improve the modified visibility method.
In this section, we consider the problem of testing the fidelity between the maximally entangled
state |(+) ih(+) | and the given state using data
(k1 , k2 , k3 ) subject to the multinomial Poisson distribu22
tion Poi( 2+1
6 t1 , 6 t2 , t3 ) with the assumption that
the parameter is unknown. In this problem, it is natural to assume that we can select the time allocation with
the constraint for the total time t1 + t2 + t3 = t.
The performance of the time allocation (t1 , t2 , t3 ) is
evaluated by the variance (34). The Fisher information
t1 t2
3
22
2+1
6 t1 + 6 t2 +t3
(43)
( 2+1
6 t1 +
2
6 t2
2
6 t2 + t3
2t1
2t2
t3 )( 3(2+1) + 3(22)
)
2 2
( t1 t
3 )
.
(44)
2t1
3(2+1)
2t2
3(22)
t t
( 1 3 2 )2
22
2+1
6 t1 + 6 t2 +t3
t1 +t3 =t 3(2 + 1)
max
VII.
2t2
3(22) )
2t1
+
( 3(2+1)
( t31 )2
2+1
6 t1
+ t3
2t
3(2 + 1)(1 +
2+1 2
6 )
(45)
(ii)
2t2
t2 +t3 =t 3(2 2)
max
( t32 )2
22
6 t2
+ t3
2t
3(2 2)(1 +
22 2
6 )
(46)
and
2t2
2t1
+
t1 +t2 =t 3(2 + 1)
3(2 2)
q
q
2+1 2
1
( 13 22
2+1 + 3
22 ) t
q
q
=
22 2
( 2+1
6 +
6 )
(iii) max
2 2
( t1 t
3 )
2+1
6 t1
6t
,
(2 + 1)(2 2)( 2 + 1 + 2 2)2
22
6 t2
(47)
using the results of (i) coincidence and total flux measurements, (ii) anti-coincidence and total flux measurements,
and (iii) coincidence and anti-coincidence measurements,
respectively. The ratio of (47) to (45) is equal to
3( 6 + 2 + 1)2
> 1,
(48)
2(2 2)( 2 + 1 + 2 2)2
as shown in Appendix B. That is, the measurement using
the coincidence and the anti-coincidence provides better
9
test than that using the coincidence and the total flux.
Hence, we compare (ii) with (iii), and obtain
max
t1 +t2 +t3 =t
2t1
2t2
+
3(2 + 1) 3(2 2)
2 2
( t1 t
3 )
22
2+1
6 t1 +
6 t2 + t3
4t
(22)( 6+ 22)2
6t
if 1 < 1
(49)
if 0 1
2(21 + 1)( 21 + 1 + 2 21 )2
= 1.
3( 6 + 2 21 )2
2(1)
22
t,
2+1+ 22
The
optimal
asymptotic
variance
is
threshold 0
is much better
than that obtained by the modified visibility method.
The ratio of the optimal asymptotic variance is given by
(50)
2(1)
t3 =
t. Otherwise, the maximum is attained
6+
2+1
t2 = 2+1+
t, t3 = 0. The
by t1 =
22
optimal time allocation shown in Fig. 1 implies that we
should measure the counts on the anti-coincidence bases
preferentially over other bases.
( 2 2 + 1 + 2)2
< 1.
6
(51)
t 22
0
t1 = 2 +1+22
, to obtain the total count n1 . Then,
0
0
we measure the count
on the anti-coincidence bases for a
20
+1
period of t2 = 2 t+1+
to obtain the total count
220
0
n2 . Note that the optimal time allocation depends on the
threshold of our hypothesis. Finally, we apply the UMP
test of of the hypothesis:
1.0
H0 : p
0.8
0.8
0.6
0.4
0.6
0.4
0.2
0.2
0.0
0.0
1.0
0.4
0.6
0.8
t 6
, t3
6+ 2(10 )
2t1
( 3((2+1)+6)
+
(2+1)
3((2+1)+6) t1
22
22+ 1+2
6+
2(10 )
.
2(10 )
1.0
If the dark count parameter is known but is not negligible, the Fisher information matrix is given by
0.899519
fidelity
versus H1 : p <
0.0
0.2
22
22+ 1+2
2t2
3((22)+6) )
(22)
3((22)+6) t2
(2+1)
(22)
3((2+1)+6) t1 3((22)+6) t2
2+1
22
1
2+1
22
(2+1)+6 6 t1 + (22)+6 6 t2 + t3
(52)
equal to
f (t1 , t2 , t3 )
2t2
2t1
+
:=(
3((2 + 1) + 6) 3((2 2) + 6)
(2+1)
t1
( 3((2+1)+6)
(2+1)
2+1
t1 +
(22)
2
3((22)+6) t2 )
).
(22) 22
t2 + t3
10
Then, we apply Lemmas 1 and 2 in Appendix A to
f (t1 ,t2 ,t3 )
2
2
with a = 3((2+1)+6)
, b = 3((22)+6)
,
(2+1) 2+1
c = (2+1)+6
6 , d =
optimized value:
(22) 22
(22)+6 6 ,
t2 +t3 =t
((2 2) +
t1 +t3 =t
((2 + 1) +
4t
q
6((2+1)+6) 2
)
4t
q
6((22)+6) 2
)
(54)
(53)
and
3
max f (t1 , t2 , 0) =t
(2+1)
3 ((2+1)+)((22)+6)
q
q
+ (2 2) 6((22)+6)
(2 + 1) 6((2+1)+6)
2
3
2 t
=
3((2 + 1) + )((2 2) + 6) 2+1
+ 22
t1 +t2 =t
((2+1)+)((22)+6)
(22)
(2+1)+6
(22)+6
62 t
=
2 .
p
p
(2 + 1) (2 2) + 6 + (2 2) (2 + 1) + 6
(55)
2
q
6((2+1)+6)
3 (2 + 1) +
2
p
p
2 (2 + 1) (2 2) + 6 + (2 2) (2 + 1) + 6
!2
p
(2 + 1) + 6((2 + 1) + 6)
3
p
p
=
> 1,
2 (2 + 1) (2 2) + 6 + (2 2) (2 + 1) + 6
where the final inequality is derived in Appendix B.
Therefore, the measurement using the coincidence and
the anti-coincidence provides better test than that using
the coincidence and the total flux, as in the case of = 0.
Define 1 and for = 6/ < 1 as
p
p
p
1 + 3 1 = 3/2
p
p
p
1 + 2 + 2 2 + = 3/2.
The parameter 1 is calculated to be 0.375. As shown in
Appendix C, the measurement using the coincidence and
the anti-coincidence provides better test than that using
the anti-coincidence and the total flux, if the fidelity is
(56)
f (t1 , t2 , t3 )
t1 +t2 +t3 =t
2
4 t
((22) + 6((22)+6))2
62 t
(2+1)
The
t1
t3
and
(22)+6+(22)
optimal
time
0,
t1
t2
if >
(2+1)+6
otherwise.
2
(57)
allocation
is
given
by
t 6((22)+6)
,
and
6((22)+6)+(22)
t(22)
for
>
6((22)+6)+(22)
(22)
t(22)
(2+1)+6
(2+1)+6+(2+1)
(22)+6
11
t2 =
(22)
t(2+1)
(22)+6
(2+1)+6+(2+1)
(22)+6
, t3 = 0
parameter is zero.
(i) Modified visibility:
(2+1)(22)
.
t
0.98
0.96
0.94
(iiia) Design II ( known), estimation from the anticoincidence count: The asymptotic variance is
3(22)
2t .
0.92
0.9
0.1
0.2
0.3 0.375
normalized dark counts
VIII.
2t2
2t1
+
).
3((2 + 1) + 6) 3((2 2) + 6)
(58) =
2 t
3((2+1)+6)
22 t
3((22)+6)
if <
if
Fig. 3 shows the comparison, where the asymptotic variances in (iia)-(iiib) are normalized by the one in (i).
The anti-coincidence measurement provides the best estimation for high ( > 0.25) fidelity. When is unknown, the measurement with the anti-coincidence count
and the coincidence count is better than that with the
anti-coincidence count and the total flux count for <
0.899519. For higher fidelity, the anti-coincidence count
and the total flux count turns to be better, but the difference is small.
(58)
1
4
1
4.
(59)
The above optimization shows that when 41 , the anticoincidence count (t1 = 0; t2 = t) is better than the coincidence count (t1 = t; t2 = 0). In fact, Barbieri et al.[12]
measured the sum of the counts on the anti-coincidence
bases |HV i, |V Hi, |DXi, |XDi, |RRi, |LLi to realize the
entanglement witness in their experiment. In this case,
the variance is 3((22)+6)
. When we observe the sum
22 t
of counting number n2 , the estimated value of is given
by 1+3( nt2 ), which is the solution of ( 22
6 +)t = n2 .
The UMP test is given from the UMP test of Poisson distribution.
IX.
(iiib) Design II ( known), estimation from the coincidence count The asymptotic variance is 3(2+1)
2t .
We compare the asymptotic variances of the following designs for time allocation, when the dark count
2
variancevariance of HiL
threshold
1.75
1.5
1.25
1
0.75
0.5
0.25
0
0.25
0.5
fidelity
0.75
0.899 1
12
X.
Optimal Allocation
(60)
( HV + + V H + + DX + + XD + + RR + + LL + )2
,
42 t
(61)
( xy + )t
txy =
,
HV + + V H + + DX + + XD + + RR + + LL +
Two-stage Method
The optimal time allocation derived above is not applicable in the experiment, because it depends on the
unknown parameters HV , V H , DX , XD , RR , and
LL . In order to resolve this problem, we introduce a
two-stage method, where the total measurement time t
is divided into tf for the first stage and ts for the sec-
(62)
mxy
(x,y)B
xy =
(x,y)B
13
ANALYSIS OF EXPERIMENTAL DATA
NLC (BBOs)
Pol
800 nm
Coinc.
400 nm
tp 100 fs
HWP
QWP
PBS
SPDM
IF
Quartz
Plate
Bereck
Compensator
0.3
0.25
Probability
XI.
0.2
0.15
0.1
0.05
0
0
10
15
20
25
30
Counts
when the visibility of the two-photon states was estimated to be 0.92. The measurement time was 40 seconds on each basis. The distribution of the coincidence
FIG. 5: Distribution of the coincidence counts obtained in
events obeys the Poisson distribution. Only small numone second on the basis |V Hi, |HV i, |XDi, |DXi, |RRi, and
bers of coincidence were observed on |HV i and |V Hi
|LLi. Bars present the histograms of the measured numbers,
bases. Those observations agree with the prediction,
and lines show the Poisson distribution with the mean values
therefore, we expect that the hypothesis testing in the
estimated from the experiment. Measurement time was 40
previous section can be applied.
seconds for each basis.
In the following, we compare four testing methods on experimental data with the fixed total time t.
The testing method employ the different time allocations
{tHH , tV V , tDD , tXX , tRL , tLR , tHV , tV H , tDX , tXD , tRR , tLL }
(ii) Design I: is unknown. The coincidence and antibetween the measurement bases:
coincidence counts are measured with the optimal
time allocation at the target threshold 0 ;
(i) Modified visibility method: is unknown. The
coincidence and the anti-coincidence are measured
with the equal time allocation;
t1
tHH = tV V = tDD = tXX = tRL = tLR =
tHH = tV V = tDD = tXX = tRL = tLR
6
t
t2
= tHV = tV H = tDX = tXD = tRR = tLL =
tHV = tV H = tDX = tXD = tRR = tLL = ,
. (64)
(65)
12
6
14
where
).
culated the p-value approximately by ( n(t/6)(22
t 2 20
t1 =
20 + 1 + 2 20
t 20 + 1
.
t2 =
20 + 1 + 2 20
(t/6)(220 )
(66)
t
,
6
(67)
tf
6
(68)
mxy
(x,y)B
txy = P
(69)
(i)
(ii)
(iii)
(iv)
p-value at 0.875 0.343 0.0736 0.0438 0.0308
15
dont need any further information of the probability distribution and the tested state. The test can be further
improved by optimizing time allocation between the anticoincidence bases, when the error from the maximally
entangled state is anisotropic. However, this time allocation requires the expectation values on the coincidence
counts, so that we need to apply the two stage method.
0.4
p-value
0.3
APPENDIX A: OPTIMIZATION OF FISHER
INFORMATION
0.2
0.1
0.872
at2
at
act21
=
ct1 + t3
( c + 1)2
xt
c1 .
Then,
act21
a
=
ct1 + t3
(c 1)2
ct2
x
+ (c + 1)t .
x
c
t
and t3 = tc+1
. Thus,
c+1
at1
0.1
0.08
p-value
(A1)
act21
t1 ,t3 0, ct1 +t3 =t
ct1 + t3
a
at
2 ct + (c + 1)t =
.
=
2
(c 1)
( c + 1)2
max
0.06
0.04
0.02
0.872 0.873 0.874 0.875 0.876
threshold for hypothesis
XII.
CONCLUSION
at1
( act1 bct2 )2
max
at1 + bt2
t1 ,t2 0, t1 +t2 =t
ct1 + dt2
2
t( ad + bc)
=
.
(A2)
( c + d)2
holds,
and this maximum value is attained when t1 =
c
t
t d
.
, t2 = c+
c+ d
d
Proof: Letting x := ct1 + dt2 , we have t1 =
t2 = xct
dc . Then,
dtx
dc
( act1 bct2 )2
at1 + bt2
ct1 + dt2
!2
ad + bc
cdt2
(c + d)t x
.
=
dc
x
and
16
Hence,
the maximum
is attained at x =
t c
t d
and t2 = . Thus,
c+ d
c+ d
cdt, i.e., t1 =
( act1 bdt2 )2
max
at1 + bt2
t1 ,t2 0, t1 +t2 =t
ct1 + dt2
!
2
t( ad + bc)2
ad + bc
=
.
(c + d)t 2 cdt =
dc
( c + d)2
Further, three-parameter case can be maximized as follows.
is
equal
to
at1 + bt2
( act1 bdt2 )2
ct1 + dt2 + t3
the
2
( act1 bdt2 )2
bdt2
,
max
at1 + bt2
.
dt2 + t3 t1 ,t2 0, t1 +t2 =t
ct1 + dt2
Proof: Define
two parameters x := ct1 + dt2 + t3 and
y := cdt1 bdt2 . Then, the range of x and y forms a
convex set. Since
bd(x t) + (d 1)y
,
t1 =
bd(c 1) + ac(d 1)
ac(x t) (c 1)y
.
t2 =
ac(c 1) + bd(d 1)
Hence,
( act1 bdt2 )2
at1 + bt2
ct + dt2 + t3
1
a bd
=
bd(c 1) + ac(d 1)
b ac
+
(x t)
ac(c 1) + bd(d 1)
a(d 1)
+
bd(c 1) + ac(d 1)
b(c 1)
y2
y
x
ac(c 1) + bd(d 1)
=
1
1
B2
(y Bx)2 + (
+ A)x At,
x
2
4
b ac
a bd
+
, B :=
bd(c1)+ ac(d1)
ac(c1)+ bd(d1)
a(d1)
b(c1)
where A :=
It is sufficient to show
r
p
3
(2 + 1) + 6((2 + 1) + 6)
2
p
(2 + 1) (2 2) + 6
p
+ (2 2) (2 + 1) + 6 > 0.
By putting :=
6
,
(B1)
LHS of (B1)
r
p
3
=
(2 + 1) + 3 (2 + 1) + )
2
p
p
(2 + 1) (2 2) + (2 2) (2 + 1) +
r
p
3
=
(2 + 1) + (2 + 1) (2 + 1) + )
2
p
(2 + 1) (2 2) +
!
r
p
3 p
+ (2 + 1) + ) (2 2) + .
=(2 + 1)
2
Since 0 1, we have
r
p
3 p
+ (2 + 1) + ) (2 2) +
2
r
+ 1 + 2 + .
2
( [0, ])
Further, the function 1 + 2 +q
17
if and only if 1 > and > .
p
3
(2 2) + 6((2 2) + 6)
2
p
(2 + 1) (2 2) + 6
p
+ (2 2) (2 + 1) + 6 > 0
6
Define
i by
(C1)
if and only if
< 1 and 6 . By putting :=
6
,
LHS of (C1)
r
p
3
(2 2) + 3 (2 2) + )
=
2
p
p
(2 + 1) (2 2) + (2 2) (2 + 1) +
r
p
3
=
(2 2) + (2 2) (2 2) +
2
p
(2 2) (2 + 1) +
!
r
p
3 p
+ (2 2) + (2 + 1) + .
=(2 2)
2
Since 0 1 and 0,
r
p
3 p
+ (2 2) + (2 + 1) + > 0
2
min D(Poi(0, . . . , 0,
i , 0, . . . , 0)kPoi(1 , . . . , m ) = R.
w =c0
In fact,
=
=
=
i=1
i i
m
X
i 0: w =c0
j=1
j a + a log
min
0,0: i +M =c0
c0
awi
wi a + a log c0
c0
wM wi
wM + a log
wM
a
i
+ a + a log
if a
if a <
c0 (wM wi )
wM wi
c0 (wM wi )
.
wM wi
pi min
D(Poi(0, . . . , 0, ai , 0, . . . , 0)kPoi(1 , . . . , m )).
w =c0
(D1)
i=1
w =c0
i=1
min
Lemma 5 implies
k
X
i=1
min
D(Poi(p1 a1 , . . . , pm am )kPoi(1 , . . . , m ))
D(Poi(0, . . . , 0, a, 0, . . . , 0)kPoi(1 , . . . , m )
k
)
) (
X k
min D(Poi(k , . . . , k )kPoi( , . . . , )) R
1
m
1
m
i
w =c0
1 .
(k1 , . . . , km
)
(k1 , . . . , km
) P
m wk <c
k
X
min
i 0: w =c0
k
X
pi min
D(Poi(0, . . . , 0, ai , 0, . . . , 0)kPoi(1 , . . . , m ))
w =c0
i=1
k
k
X
X
pi m,i ))
pi 1,i , . . . ,
D(Poi(p1 a1 , . . . , pm am )kPoi(
i=1
i=1
min
D(Poi(p1 a1 , . . . , pm am )kPoi(1 , . . . , m )).
w =c0
We
choose
1,i , . . . , m,i
such
that
minw =c0 D(Poi(0, . . . , 0, ai , 0, . . . , 0)kPoi(1 , . . . , m )) =
D(Poi(0, . . . , 0, ai , 0, . . . , 0)kPoi(1,i , . . . , m,i )). Then,
18
Hence,
satisfy
m
m
X
X
i
i log )
p( (i i ) +
i
i=1
i=1
m
m
X
X
i log i )
+ (1 p)( (i i ) +
i
i=1
i=1
m
X
i=1
m
X
i=1
(pi + (1 p)i )
.
(pi + (1 p)i )
p(
i=1
m
X
i=1
i log
m
X
i log i )
) + (1 p)(
i
i
i=1
(pi + (1 p)i )
.
(pi + (1 p)i )
m
m
X
X
i
i log i )
i log ) + (1 p)(
p(
i
i
i=1
i=1
m
X
pi
i
) log( )
(p
+
(1
p)
i
i
i
i=1
(1 p)i
log( i )
+
(pi + (1 p)i )
i
m
X
(pi + (1 p)i )
(pi + (1 p)i ) log
.
(pi + (1 p)i )
i=1
(pi + (1 p)i )
(pi + (1 p)i )
)
(pi + (1 p)i )
i
i
(1 p)i
pi
+
)
= log(
(pi + (1 p)i ) i
(pi + (1 p)i ) i
i
pi
log( )
(pi + (1 p)i )
i
(1 p)i
i
+
log(
).
(pi + (1 p)i )
i
log(