Sen, P.K. and Ghosh, Maloy; (1970)Sequential confidence interval for the regression coefficient based on Kendall's Tau."

SEQUENTIAL CONFIDENCE INTERVAL FOR TIlE
REGRESSION COEFFICIENT BASED ON KENDALL'S TAU
By
Malay Ghosh and Pranab Kumar Sen
Department of Biostatistics
University of North Carolina at Chapel Hill
Institute of Statistics Mimeo Series No. 658
January 1970
Sequential Confidence Interval for the Regression Coefficient
Based on Kendall's Tau*
BY MALAY GHOSH AND PRANAB KUMAR SEN
Department of Biostatistics
University of North Carolina at Chapel Hill
SUMMARY
The object of the present investigation is to consider a robust procedure
for the problem of providing a bounded-length confidence interval for the
regression coefficient (in a simple regression model) based on Kendall's (1955)
tau.
The problem of estimating the difference in the location parameters in the
two-sample case may be viewed as a special case of our problem.
It is shown that
the estimate of the regression coefficient based on Kendall's tau [cf. Sen (1968)],
as extended here in the sequential case, possesses certain desirable properties.
Comparison with the procedure based on the least squares estimator [considered
by GIeser (1965) and Albert (1966)] is also made.
1.
INTRODUCTION
Consider a sequence {Xl,X Z, .... } of independent (real valued) random
variables with (absolutely) continuous cumulative distribution functions (cdf)
F (x) ,F 2 (x) , ... , where
l
F.(x)
= F(x-a-8t.),
1
1
i=1,2, ... ;
the t.1 are known regression constants, 8 is the regression coefficient and a
is a nuisance parameter.
It is desired to determine a confidence interval
* Work supported by the National Institutes of Health, Grant GM-12868.
(1.1)
2
I n = {s:
SL ,n -< S -< ~-
} such that (i) P{S£I n } = I-a, the desired confidence
coefficient and (ii) 0 -< ~- ~L ,n -< 2d, (d predetermined). Since the form
~U,n
~u,n
F( ) is not known, we are not in a position to prescribe any fixed-sample size
procedure valid for all F.
Sequential procedures for such a problem, based on the classical least
squares estimators of a and S, are due to GIeser (1965) and Albert (1966).
These procedures, like ours (to follow), are based on the method suggested by
Anscombe (1952) and Chow and Robbins (1965).
However, being based on the least
squares estimates, these procedures are vulnerable to gross errors or outliers,
and may be quite inefficient for distributions with heavy tails (e.g., the Cauchy,
logistic or the double exponential distribution).
For this reason, we consider
here an alternative robust procedure based on the nonparametric estimate of S,
considered in Sen (1968).
The procedure is explained in Section 2.
deals with the main results of the paper.
Section 3
Section 4 is devoted to the study of
the asymptotic relative efficiency (A.R.E.) of the different procedures.
In
Section 5, we consider the two-sample location problem, which is a particular
case of (1.1) when the t.1 can be either 0 or 1.
The appendix is devoted to the
asymptotic linearity of a stochastic process involving Kendall's tau; this result
is used repeatedly in Section 3.
2.
THE PROPOSED SEQUENTIAL PROCEDURE
For every real b(-oo<b<oo), define Z.(b)
= X.-bt.,
i=I,2, ... , and let
1
1
1
Un(b) = (n)-1
L c(t.-t.)c(Z.(b)-Z.(b)),
2
..
J 1
J
1
1~l<J~n
where c(u) is equal to 1, 0 or -1 according as u>, = or <0.
(2.1)
Note that by defini-
tion,Un(S) is distribution-free and its distribution (symmetric about zero) is
3
tabulated (for small values of n) by Kendall (1955) and Smid (1956).
If we let
a
V = (1/18){n(n-l)(2n+5)-L:- nl u .(u .-1)(2u .+5)},
n
J= nJ nJ
nJ
(2.2)
[where among (tl, ... ,t) there are a (>2) distinct values with frequencies u "
n
n~
j=l, ... ,an]' then for large n, (~)Un(8)~ has approximately a nonna1 distribution with mean
a and
variance 1 [cf. Hoeffding (1948) and Kendall (1955)].
such, for each nand !n = (tl, ... ,tn), we can find an
U~
As
t such that
'-n
P{-U*n t -< Un (8) -< U*n t 18} = I-an'
'-n
'-n
where
(2.3)
an~ as n~. [For large n, U~'!n ~ \~./2~/(~), where \]./2 is the upper
100a/2% point of the standard normal distribution.]
As in Sen (1968), we use
(2.3) to derive the following (fixed-sample) confidence interval for 8.
y..
1J
where Sn = {(i,j): t·ft.,
J
1
= (X. -X. ) /
J
1
l~i<j~n}.
(t . - t . ),
J
1
(i, j) E:S ,
n
Let
(2.4)
We denote by n* the mnnber of elements of
Sn (i.e., the number of distinct pairs (t.,t.)
in -n
t ), and denote the n* ordered
1
]
values in (2.4) by
(2.5)
Also, let
M(l) = h[n*-(n)U*
]
n
2
2 nt'
'-n
(2.6)
Then, as in Sen (1968), we have
p{y
(1)
n(Mn )
< 8 < Y
(2)
n(Mn +1)
IS} = I-an .
(2.7)
4
Now, in order to obtain a confidence interval for S of length
~
2d, we define a
n0 (the initial sample size,
stopping variable N=N(d) to be the first integer>
may be 2 or more), for which Y
(2)
-y
(1))
n(Mn +1) n(Mn
(sequential) confidence interval for S is
~
2d.
Then, our proposed
(2.8)
In principle, our procedure is similar to that of Anscombe (1952) and of Chow and
Robbins (1965).
However, they used the sample mean square due to error to set
up an appropriate confidence interval, whereas we use Kendall's tau to derive such
an interval.
The main results are stated in the next section.
PROPERTIES OF TIlE CONFIDENCE INTERVAL IN(d)
3.
In the remainder of the paper, we shall stick to the following notations and
assumptions.
Let
(3.1)
. n1 (t.1 -t-n )( 1.
n = {L1=
p
n+l
- ---,,-)}
~.
I {AnTn };
(3.2)
a
An2 = (1/12){n(n 2 -l) - I.nlu
.(u 2 .-l)}.
J= nJ nJ
(3.3)
Then, concerning {t }, we assume that
-n
max
-
(i) 1·
<l<n It"1-tn IIT n+0 as n-+oo,
(3.4)
(3.5)
n
(1·1·1·) limn inf[,a
L" lU . (n-u. )1 n 2] >0 ,
J= nJ
nJ
(3.6)
5
(iv)
(v)
1·
n~ Pn = P, where Ipl>o,
(3.7)
r n2 = Q(n) is a strictly increasing function of n, such that for
every a>O
1·
lim
n: Q(ann)/Q(n) = sea) whenever n~ an = a,
(3.8)
where sea) is strictly monotonically increasing and continuous
in a, and s(l)=l.
Note that our (ii) and (v) are less restrictive than similar conditions of
GIeser (1965), who requires that lim n- l T2 = c>O. If t.=i, i=1,2, •.. , (i)-(v)
n~
hold, but GIeser's does not.
n
1
Note that by (3.3) and (3.6), we have
(3.9)
Concerning F(x) in (1.1), we assume that f(x) = FI(x) exists (a.e.), and is
continuous in x (a.e.); also assume that
00
Jf 2 (x)dx = B(F)<oo.
(3.10)
-00
Our main theorem of the paper is the following.
Theorem 1.
Under the assumptions made above, N(=N(d)) is a non-increasing function
of d(>O) , N(d) is finite almost surely (a.s.), E[N(d)]<oo for all d>O, lim N(d)=oo
d+O
a.s. , and lim E[N(d)]=oo. Further,
-;....;....<:...-- d+O
lim N(d)/Q-l(v(d)) = 1 a.s.,
d+O
(3.11)
for all F,
(3.12)
lim E[N(d)]/Q-l(v(d)) = 1,
d+O
(3.13)
lim P{S£IN(d)} = I-a
d+O
6
where
(3.14)
Proof.
~
L,n
The proof of the theorem is completed in several steps.
= Yn(M(l)) and
n
~~u,n
First, we let
= Yn(M(2)+1)' which are defined as in (2.6) and (2.7).
n
Then we have the following.
For every 0>0 there exists an no (0), such that for n>n
(0),
- 0
Lennna 3.1.
(3.15)
P{SL
,n
<
S-(log n)/(p T )} < n-(l+o).
n n -
(3.16)
We shall only prove (3.15) as (3.16) follows on the same line.
tion of ~~u
,n
,U
n
P{~.
~u,n
(Bu ,n) > -U~ '-n
t.
By defini-
Hence,
> S+(log n)/(p nTn)} = P{Un (S+(log n)/p nTn) -> -U*n,! }
= P{Un (S+(log n)/pnn
T ) - E[Un (S+(log n)/pnn
T )]
~
-
U~,t
-n
-E[Un(S+(log n)/PnTn)]}·
(3.17)
-~
.?2 n
Using the fact that for large n, U~,t ~ Ta/2v~/(2) = O(n ), and
-n
E[Un(S+(log n)/PnTn)] = - 4An (log n)(2)
write -U~,t -E[Un(S+(log n)/PnTn)] as
-1
[B(F)+o(l)] = O(n
-~
log n), we may
-n
n -1
4An (log n)(2)
B(F) [l+o(l)+O(log n)
-1
].
Now, we make use of a result by Hoeffding (1963; (5.7)) on the deviation of a
(3.18)
7
U-statistic and obtain from (3.17) and (3.18) that for n adequately large
n
A
2
P{£L
> S+(log n)/p nT}
~U,n
n -< exp[-[-2]t n/4]
< exp[-(n-2)tn2 /8],
(3.19)
n -1
-k
where t n = 4An (10g n)(2) B(F) = O(n 2 log n). Thus, (n-2)t~/8 can be made
greater than (l+o)log n, for any 0>0, when n is made greater than no(o). Hence,
for n>n
(0), the right hand side of (3.19) can be made less than
-0
exp[-(l+o)log n] = n-(l+o).
Q.E.D.
Lemma 3.2.
There exists two positive numbers °1 and ° 2 and an n, say no (01,02)'
such that for all n ~ no (01,02)'
-0
p{11.! B(F)PnTn(SU,n-SL,n)/Ta/2-ll
~ O(n llog n)}
-0 -1
<
O(n 2 ).
(3.20)
The proof of the lemma directly follows from lemma 3.1 and the following
theorem whose proof is sketched in the Appendix.
Theorem 3.3.
There exists two positive °1'° 2 and an no(01'02) such that for
n ~ no (01,02)'
P
P{I a ISul
< og n I (n2){Un (S+a/p nTn ) - Un (S)} + 4aB(F)An I~n
~
Lemma 3.4.
O(n
-0 1
log n)} .:. O(n
-1-0 2
).
(3.21)
For every real x (-oo<x<oo)
(3.22)
-00
8
The proof is given in Sen (1968).
For the asymptotic normality of
PNTN(SU ,N- S), we require, as in Anscombe (1952), the 'uniform continuity in
probability" of {it.
"'u,n} with respect to pT.
n n
lerrnna.
Lerrnna 3.5.
For every positive
E:
For this, we have the following
and n there exists a 0(>0), such that as
n+oo
(3.23 )
and a similar statement holds for
Proof.
{6L,n} .
By virtue of lemma 3.1 and theorem 3.3, and the definition of ~T
...u ,n'
= {Ta /2/ 1IZ B(F)}{(PnTn)/(Pn,Tn,)-l} +
[1IZB(F)]-l{[(PnTn)/(Pn,Tn ,)]
(~')/~,Un,(S)-(~)V~~n(S)}+O(l)
+ [(n2)/~]
[Un ,(S)-Un (S)] + 0(1).
n
(3.24)
Since Un(S) is a U-statistic in the independent and identically distributed
random variables Zi(S),
l~i~n,
a reverse martingale sequence.
it follows from Berk (1966) that {Un(S)} forms
Hence, if we write n =no, and let W. = U + . (S) nno-J
o
J
U + ·+l(S), 1~<2n ,W = U
(S), it follows that {WO,Wl , ... ,W2 } possesses
n no -J
- 0 0
n+no
no
the properties of amartingale difference sequence. Hence, by using the well-known
9
Kolmogorov inequality [cf.
Lo~ve
(1963, p. 386)], we have
max
P{l<1<
2n IWl+···+W.J I>tn} -< t n-2 E(Wl +... +W2n )2
~- 0
0
°
(3.25)
But, WI+... +WJo = Un+n - JO Cfn -Un+no ((3), 19<- 2n0 , and the right hand side of (3.25)
o
is equal to t~2{V[Un_n (S)]-V[Un+n (S)]} = (4/9)(2not~2/(n2-n~))(1+O(n-l)).
9J.
Hence, letting t n = nn ~, we have
0
(3.26)
This limnediately leads to (by proper choice of 6)
(3.27)
for n sufficiently large.
Also, by (5.7) of Hoeffding (1963),
p{IUn (S)I>K£ ,,}<£",
where K£ ,,<00.
-
(3.28)
[s(1-6), s(1+6)], and hence, differs from 1 by an arbitrarily small quantity, as
(by (iv) and (v)), n(~)/Vn = 9/2 + O(n-l) (by (iii)), I (~')-(~)1~6(2+6)n2,
3
3
3
2
and IV-~-V-~I
n' n = IVn ,-Vn I/{I~,+~I~,}
n n n n = 0(n- )Ivn ,-Vn I = 0(n- ) [(2/9){n (1+36+0(6 ))n 3 } + 0(6)n 3 ] = 0(6), where the last order follows by using (iii) on the second
ll+oo
term of the right hand side of (2.2).
(3.28) and the above discussion.
Thus, (3.23) follows from (3.24), (3.27),
Hence the lennna.
Returning now to the proof of theorem 1, we note that by lennna 3. 2 and the
definition of N(d) , for all d>O, N(d) is finite a.s., and it is non-increasing
in d.
To show that E[N(d)]<oo for every d>O, we write
10
E[N(d)] = LnP{N(d)=n} = Ln> OP{N(d»n}.
(3.29)
Hence, it suffices to show that for every d>O, Ln>OP{N(d»n} converges; a
sufficient condition for this is to show that as
ll+OO,
P{N(d»n} ~ O(n- l - y), where y>O.
(3.30)
Now, by definition
< P{(~-8 »2d} = P{I.rB(F)p nTn (~-6 »2dl.r B(F)pnTn}
""U ,n L,n
""U, n L,n
-1-0 2
~ O(n
), O2>0,
-
(3.31)
-0
as by (3.20), I.r B(F)p nTn (~_
-8 )-1 = O(n 1 log n) with probability ~ l-O(n
""u ,n L ,n
and by (ii) and (iv) PnTn-t<Jo as ll+OO. Thus, E[N(d)]<oo, for every d>O. Using the
fact that~.
""u ,n
-SL ,n>0
(a.s.) for each n, we have lim N(d) =
d~O
00.
-1-0
Now, from the
Monotone Convergence Theorem, lim E[N(d)]=oo.
d~O
(3.11) is a direct consequence of (3.20), the definition of v(d) in (3.14),
(iv) in (3.7) and (v) after (3.7).
(3.12) follows readily from theorem 1 of
Anscombe (1952) along with our 1errnna 3.2, 1errnna 3.4 and 1errnna 3.5.
To prove
(3.13), we write
E[N(d)]/Q
-1
[v(d)] = {Q
-1-1
[v(d)]} Ll +L 2+L 3 nP{N(d)=n},
where the summation Ll extends over all n<nl(d) , L2 over all
L3 over all n>n 2 (d) , and where
nl(d)<n~n2(d)
Q(n.(d)) = v(d){l+(-l)i€ + (1+(-1)i)/2}, i=1,2.
1
(3.32)
and
(3.33)
2),
11
Since llin v(d) = and llin N(d)/Q-l(v(d)) = 1 a.s., and Q( ) is ,for every
d-+O
d-+O
£>0, there exists a value of d, say d (>0), such that for all O<d<d ,
00
£
-£
P{nl(d) ~ N(d) ~ nZ(d)} ~ P{IN(d)/Q-l(v(d))-ll<£} ~ l-n, where n is arbitrarily
small.
Hence, for d<d ,
-£
{Q
-1
[v(d)]}
-1
LlnP{N(d)=n}
~
(l-£)P{N(d)<nl(d)}
~
n(l-£).
(3.34 )
Also, proceeding along the same line as in (3.31), we have
{Q-1[V(d)]}-l L3nP{N(d)=n} = {Q-1(v(d))}-lnz (d)P{N(d)=n (d)}
z
+
{Q-l(v(d))}-l L3P{N(d»n}
~
{Q
-1
(v(d))}
-1
[L 3P{N(d»n}
+
n Z(d)P{N(d»nZ(d)-l}]
-8
= {Q-1[v(d)]} -l[o(n (d))
z
-8
= O[(nZ(d))
-1-8
Z + O(nZ(d))O([nZ(d)]
Z)]
Z
] -+ 0 as d-+O,
-1
since Q [v(d)]+oo as d-+O and hence, nz(d)+oo as d-+O.
(3.35)
Finally, elementary computa-
tions yield that
(3.36)
Hence, (3.13) follows from (3.34), (3.35) and (3.36).
4.
Q.E.D.
ASYMPTOTIC RELATIVE EFFICIENCY
For two procedures A and B for determining (sequentially) bounded-length
confidence intervals for S (with the same bound Zd), we define the A.R.E. of
A with respect to B as
12
where NA(d) and NB(d) stand for the respective stopping variables.
It may be noted that though GIeser (1965) considered the case where
n-lT~~>O as n~, his results can be easily extended to the case where our (ii)
and (v) of section 3 hold.
As such, if NL(d) stands for the stopping variable
for his procedure (based on the least squares estimators), it follows that our
theorem 1 also holds for NL(d) with the only change that v(d) has to be replaced
by vL(d) =
02T~/2/d2,
where 02 is the variance of the distribution F(x) in (1.1).
Hence, using the result in (3.13) for both the cases and writing NK(d) for the
proposed procedure, we have
eK L = lim {Q
,
d-+O
-1
(vL(d))/Q
-1
(4.2)
(v(d))}.
Note that, by definition vL(d)/v(d) = l20 2p 2B2 (F) = e, is independent of d, and
let us write e = s(e*), so that e* = s-l(e) is monotonic in e, with e*=l when
e=l.
-1
Also, let ¢d = Q (vL(d))/Q
-1
-1
(v(d)), Q (vL(d))
= vt(d) and Q-1 (v(d)) = v*(d).
Note that by (v) of section 3, vt(d) and v*(d) both tend to
00
as d-+O.
Hence, we
have
s (e*) = e = vL (d) / v (d) = 1im {vL(d) / v (d) }
d-+O
= lim {Q(vt(d))/Q(v*(d))}
d-+O
= lim {Q(¢(d)v*(d))/Q(v*(d))}.
(4.3)
d-+O
Using (3.8) and proving by contradiction, it follows that
1im ¢ (d) = e* = s
d-+O
-1
(e) .
(4.4)
13
Hence, from (4.2) and (4.4), it follows that
=s
e K, L
-1
(4.5)
(e)
-00
In Sen (1968), various bounds for e = l2a 2p2(
f
00
f 2 (x)dx)2 are studied, with
-00
special reference to the bounds for p2.
As we shall see in the next section, for
the two-sample problem, p=l, and see) = e, so that (4.5) reduces to l2a
2
f
(
00
f2(x)dx)2,
-00
the usual A.R.E. of the Wilcoxon two-sample test with respect to the Student's
00
t-test.
Since l2a
2
(
f
f2(x)dx)2
~
0.864 for all continuous F, it follows that
-00
(4.6)
i=1,2, ... , so
Consider now the case of equispaced regression line, where t.=i,
1
that Tn2 = n(n 2-l)/12 • In this case, we get at once, that p=l and
00
eK L = (12a 2 (
,
f
1
_
f2(x)dx)2)3
(4.7)
-00
For the case of normal F, this reduces to (0.95)1/3 ~ .985, while the infimum
in (4.6) is given by (0.864)1/3 ~ .953.
This clearly indicates the robustness
and efficiency of the proposed procedure.
For many non-normal cdf (e.g.,
double exponential, logistic, Cauchy, etc.) the proposed procedure is more
efficient than the least squares procedure.
5.
lWO-SAMPLE LOCATION PROBLEM
Consider the special design, where the t.1 can either be 1 or O.
Thus, we
have only two different distributions F(x-a) and F(x-a-S), where S denotes the
difference in the location parameters of the two distributions.
If at the n-th
stage, we have mn of the t equal to 1 and the rest equal to zero, we obtain that
i
14
r n2 = mn (n-mn )/n
< n/4, for all n>l.
-
(5.1)
Looking at the definition of r n2 = Q(n) , (3.11) and (3.13), we observe that an
optimum choice of mn is [~], the integral part of~. Thus, among all designs
for obtaining a bounded-length confidence interval for S, in this problem, an
optimum design (which minimizes the expected value of N(d) for small d) consists
in taking every alternative observations for the two distributions. Here also,
l
p =p=l, and n- r 2 + 1/4, and hence, by (4.5), the A.R.E. reduces to l2a 2 ( ff 2 (x)dx)2,
n
n
00
-00
various bounds for which are well-known.
Looking at (2.4)-(2.7), we observe that
1
n* = m (n-m ) tV ifl2,
and U* t can be computed, for small values of n, from the
n
n
n'-n
extensive tables given in Owen (1963). For large n, note that Vn , defined by
(2.2), reduces to (1/18){2[n3-m~-(n-mn)3] + 3[n2-m~-(n-mn)2]} ~ {2 n 3+O(n 2),
and hence,
U~
t
can be computed by reference to the usual nonnal probability
'-n
tables.
6.
APPENDIX:
PROOF OF TIffiOREM 3.3
We may assume without any loss of generality that S=O (as otherwise, we
may set b=S+b' and work with b').
Also, we shall explicitly consider the case
of a£[O, log n] as the case of a£[-log n, 0] follows on the same line.
b
n
tV
n
Let
°1 (as n+oo), 01>0, and let n
(r/bn )log n, r=l, ... ,bn . Since, Un(b)+
r,n =
in b, for a£[n r- 1 ,n ,nr ,n], we have
(6.1)
and similar inequalities involving their expectations.
few simple steps,
Then, we get after a
15
ae[n
Sup n ] IUn (0) - Un (I
n
n + EUn (I
a PnT)
a PnT)
r-l,n' r,n
I
(6.2)
where W .
n,]
= Un (O)-Un(n.
Ip T )
] ,n n n
,n=0, EUn (0)=0).
nO
+ EU (n. Ip T ),
n ] ,n n n
O~<b
- n
,1<r<b (note that
- - n
Hence,
(6.3)
<
max
O~~bn
[(n)/..f2]
2' n
Iwn,j I
+
Using (3.4), (3.11), nj ,n-n j - 1 ,n
= log n/bn and
A~/Vn
+
3/4 as n+oo [cf. Sen
(1968)], we get,
1
(6.4)
= (n.],n-no] -,n
1 )(A JV:)[B(F)+o(I)]
n n
-6
= O(n \og n), for all j=1,2, ... ,b .
n
Let
Y ..
nl]
</>(Yn
••l,Y]nk
]·)
12.irk2.n,
1~<b
l~<bn'
= X.-(n.
Ip T )t.(l<i<n, l<i<b),
1
],n n n 1 - ~- n
= c(tk-t.)
[c(Yn]
k'- Ynl]
k'- Ynl]
.. )],
. ·+(n·],nIp nTn )(tk-t.))-c(Y
1
1
n]
.
Wrlte,
W . = n -1 (n-l) -1 L\'~
[</>(Y .. ,Y k·)-E</>(Y .. ,Ynk ·) ] ,
n,]
l<i k<n
nl] n]
nl]
]
Note that W . is a U-statistTc minus its expectation (see
n,]
(W 0=0).
n,
e.g., Hoeffding (1948)).
- n
-1 \' n
2n]. = n
L' 12 ., = (n-l)
1= nl]
~
We can write, W .
n,]
-1 \' n
. = E 2n]., 1~ -<b.
Then,
n
n]
=
Lk 1(~·)c(tk-t.)
= r1
1
2(2 .-~ .) + R ., where,
n] n]
n]
[F(X.-a)-F(X.-a-(n·
1
1
],nIp nTn )(tk-t.))];
1
16
1
P[\Wn,J. I [(n2)/~]
n > cln
< P[
-
Izn,J.-~ nJ. I
-0 1
log n]
> ~t ] + P[ IR . \ > ~t ],
n
nJ
(6.5)
n
-~-o
-0
where cl(>O) is not depending on n, t n = cl~(~)-l n 1 log n = O(n
llog n),
as n+oo, since, Vn = O(n 3 ) , (as n+oo), (from (3.9) and the fact that An2 /Vn ~ 3/4
as n+oo).
Also, from (3.4), (3.9) and (3.10), it follows after some algebraic
manipulations that ~ . = 4n. (A /[n(n-l)]) (B(F)+o(l)) = O(n-~log n), uniformly
nJ
J,n n
in j=1,2, ... ,bn . Hence, there exists a positive integer n l such that for n~nl'
O<t'<~
.<l-~nJ.<1, uniformly in j=1,2, ... ,bn (t'-slt).
Also, Z . is the average
n nJ
n n
nJ
of n independent random variables each assuming the values 0 and 2.
Now, using
the inequality (2.1) of Hoeffding (1963), we get, for n>nl ,
P [k
2
< [(~'./(~'.+t')}
nJ
nJ n
IZnJ. -~ nJ. I>t'n]
n(~'.+t')
nJ
n {(l-~' .)/(l-~'.-t')}
nJ
nJ n
.-t')
nJ n]
n(l-~'
(6.6)
+ [{~'./(~' .-t')}
nJ
nJ n
where
~'.
nJ
=
~
..
nJ
.-t')
n(l-~' .+t')
nJ n {(l-~' .)/(l-~'.+t')}
nJ n]
nJ
nJ n
n(~'
It follows after some algebraic simplifications that as n+oo,
nt'2
= - nt' 2 j{ 2 (~ , . ±t ' ) (l-~ , . +t ' ) } + 0 (2 (~I n. ±t I)) .
n
nJ n
nJ n
nJ n
2
-c /4
= O(log n 1 ).
Also,
(6.7)
17
(6.8)
But,
2
E(RnJo) =
n2(n~1)2
~ n2(n~1)2
n2(~-1)2
LL
1~i1k~n
LL
1<i1k~n
V[¢(Y. "Ynk°)]
nlJ
J
E[E{¢(Y .,Y k·)IY °.}-E¢(Y oo'Ynk °)]2
nlJ n J nlJ
nlJ
J
0
L L V[¢(Y.
"Ynk°)]
1~i1k.2.n
nlJ
J
< (4 1)
- n n-
~ nJ.
(6.9)
5 2
= O(n- / log n).
It follows now from (6.5)-(6.9) and the Bonferroni inequality that
b
-c 2 /4
n
P{l~b Iw ·I>t } < b O(n 1 ) + L E(R2.)/(~t2)
.9_ n nJ n - n
j=l
nJ
n
(6.10)
First make a proper choice 01 and subsequently of c1 (>0) to make the right hand
-1-8
side of (6.10) less than or equal to [(const)(n
2)] for any given 02>0. The
result now follows from (6.3), (6.4) and the Bore1-Cante11i lemma.
18
REFERENCES
ALBERT, A. (1966). Fixed size confidence ellipsoids for linear regression
parameters. Ann. Math. Statist., 37, 1602-1630.
ANSCOMBE, F. J. (1952). Large-sample theory of sequential estimation.
Proc. Camb. Phil. Soc., 48, 600-607.
BERK, R. H. (1966). Limiting behavior of posterior distributions when the
model is incorrect. Ann. Math. Statist., 37, 51-58.
CHOW, Y. S. AND ROBBINS, H. (1965). On the asymptotic theory of fixed-width
sequential confidence intervals for the mean. Ann. Math. Statist., 36,
457-462.
GLESER, L. J. (1965). On the asymptotic theory of fixed-size sequential confidence bounds for linear regression parameters. Ann. Math. Statist.,
36, 463-467.
HOEFFDING, W. (1948). A class of statistics with asymptotically normal
distribution. Ann. Math. Statist., 19, 293-325.
HOEFFDING, W. (1963). Probability inequalities for sums of bounded random
variables. J. Amer. Statist. Assoc., 58, 13-29.
KENDALL, M. G. (1955).
Company: London.
LOEvE, M. (1963).
Rank Correlation Methods.
Second edition, 1955.
Charles Griffin and
Probability Theory (3rd edition) Van Nostrand, Princeton.
OWEN, D. B. (1962). Handbook of Statistical Tables.
Reading, Massachusetts.
Addison-Wesley,
SEN, P. K. (1968). Estimates of the regression coefficient based on Kendall's
tau. J. Amer. Statist. Assoc., 63, 1379-1389.
SMID, L. J. (1956). On the distribution of the test statistics of Kendall and
Wilcoxon's two sample test when ties are present. Stat. Neerlandica, 10,
205-214.