Sen. P. K.; (1979)On Nonparametric Sequential Point Estimation of Location Bason on General Rank Order Statistics."

.t~.
ON NONPARAMETRIC SEQUENTIAL POINT ESTIMATION OF LOCATION
BASED ON GENERAL RANK ORDER STATISTICS*
by
Pranab Kumar Sen
Department of Biostatistics
University of North Carolina at Chapel Hill
Institute of Statistics Mimeo Series No. 1234
June 1979
,
ON NONPARAMETRIC SEQUENTIAL POINT ESTIMATION OF LOCATION
BASED ON GENERAL RANK ORDER STATISTICS*
PRANAB KUMAR SEN
University of· North Carolina, Chapel Hill, N.C.
ABSTRACT
A nonparametric sequential procedure for the point estimation of
location of an unspecified (symmetric) distribution based on a general
class of one-sample (signed) rank order statistics is considered and
its asymptotic theory is developed.
The asymptotic risk-efficiency of
the proposed procedure is established and certain almost sure and
moment convergence results on rank based estimators are studied in
this context.
AMS Subjeat CZassifiaation Nos:
62L12, 62L15, 62G05.
Key Words & Phrases: Almost sure representation, asymptotic normality,
asymptotic risk-efficiency, loss function, moment-convergence, rank
estimates, risk function, sequential point estimation, signed r'ank
statistics, Skorokhod-Strassen embedding, stopping variable.
*Work supported by the National Heart, Lung and Blood Instutute,
Contract No. NIH-NHLBI-71-2243 from the National Institutes of Health.
-2-
1.
INTRODUCTION
Sequential point estimation of the mean of a normal distribution
and its asymptotia pisk-effiaienay have been considered by Robbins
(1959), Starr (1966), Starr and Woodroofe (1969) and Woodroofe (1977),
among others.
Recently, Ghosh and Mukhopadhyay (1979) have studied an
asymptotically risk-efficient sequential procedure for point estimation
of the mean of an unspecified distribution (having a finite eighth
order moment).
For the dual problem of bounded-length aonfidence
intepval for the mean of an unspecified distribution, Chow and Robbins
(1965) have developed an elegant sequential procedure which is asymptotically both consistent and efficient.
Sen and Ghosh (1971) have emp1oyed_
a general class of one-sample rank order statistics for nonpapametpic
sequential intepval estimation of location.
for the point estimation problem.
Nothing parallel is done
The asymptotic theory of nonparametric
sequential point estimation of location, based on a general class of
rank statistics and derived estimates, is developed in the current paper.
The asymptotic risk-efficiency of the proposed procedure is established
and a comparison is made with other procedures.
Along with the preliminary notions, the proposed sequential procedure for the point estimation of location is formulated in Section 2.
Certain moment-convergence results on rank order estimates of location
are also studied in this section.
These results have an important role
to play in the development of the asymptotic theory in the subsequent
sections.
The main theorems of the paper are presented in Section 3 and
their proofs are considered in Section 4.
Unlike the case of Woodroofe
(1977), our stopping vaPiable does not involve a sum of independent
random variables (or martingales/reverse martingales), and hence, a
-3-
different approach is needed here.
The concluding section deals with
the asymptotic relative performances of the proposed and some other
competing procedures and the superiority of the nonparametric appraoch
is stressed in this context.
2.
THE PROPOSED PROCEDURE
{X., i ~1}
be a sequence of independent and identically
l.
distributed random variables (i.i.d.rv) with a continuous distribution
Let
function (df)
Fa(X),
x EE = (_00, (0), a real (unknown),
Fa
unknown
and where
Fa (x) =F(x - a) ,
F symmetric about O.
For nonparametric estimation of
~* ={~*(u),
0 <u <I}
+ ~* (1 -u) =0, V 0 <u < 1),
saOl'e
funation~
n(~
a, we proceed as follows.
Let
be a non-decreasing, skew-symmetric (i.e.,
~* (u)
~
(2.1)
non-constant and square-integrable
= {ep(u) = ~* ((1 +u)/2),
0 <u < I}
and, for every
1), let
(2.2)
where Un. 1· :;:; ••• :;:; Unn are the ordered rv's of a sample of size n
from the uniform (0, 1) df. Further, let "'Il
X =(Xl, ... ,X),
1 =(l, ... ,n)
n
"'Il
and for every real
b, let
X (b) =X"'I1 -bl"'I1 .
"'I1
Consider then the usual
one-sample rank order statistics
Sn (b) =S(X"'I1 (b)} =I~l.='
lsgn(x.l. -b)an (R+.
nl. (b)), bEE, n ~l,
where
for
is the rank of IXl.' -bl among
R+.(b)
nl.
i =1, ... ,n. Then, for every n(~ 1),
S (b) is \
n
.
IX -bl, ... ,lxn -bl,
I
in b: _00 < b <00 ,
while, under H : a =0 and (2.1), S (0)
O
n
symmetric about 0 and for large n,
has a distribution
-4!,;
~V N(O~
n-2g n (O)/An
~
1)
(2.5)
where
(2.6)
We define
e~l) =sup{b: Sn(b) >O},
8~2) =inf{b: Sn(b) <OJ;
(2.7)
@ =(6(1) +8(2))/2.
n
Then~
n
(2.8)
n
is a translation invariant~ median-unbiased~ robust and
en
consistent estimator of
6~
as
and~
n
-+OO~
(2.9)
where
2/B
2 and B = B(4)~ F) = food
v2=A
dx 4>* (F(x) )dF(x)
(> 0) .
(2.10)
_00
We let
v~
We shall see
= E(6n _6)2
Theorem 2.1] that
[viz.~
mild regularity conditions and
(2.11)
v~ exists «
00)
under very
Theorem 2.2] under additional
[viz.~
conditions~
nV
2
n
-+
v2
as
n -+ 00
To motivate and propose the sequential
the Zoss incurred in estimating
8
'"
Ln (c)=c
,..,
l (8n -6)
where
c
by
2
en
(2.12)
procedure~
suppose that
is
+c 2n; ,..,
c=(c1~
c2»Q.~
.-
(cost per unit sample) are specified constants.
2
the risk (for a given £) is
1
and c
(2.13)
2
=EL n (£) =clvn +c 2n
and we like to minimize this risk by a proper choice of n.
\t(£)
of(2.12) and
n =no (c)
,.., ~
(2.14)~
where
if
c /c
2 l
is
sma11~
then
Then
(2.14)
By virtue
An (c)
,.., is minimized at
-5-
(2.15)
here,
aCE) ""b(£)
means that
a(£)/b(£) -+1
as
c 2/c 1 -+O.
Thus, even
in this asymptotic setup, the minimum risk estimation of e demands
the know1dege of v,
which, by (2.10), depends on the unknown df
B($, F)).
(through the functional
sequential
pro~edure
F
Hence, we take recourse to a
based on a sequential estimator of
v.
As has been mentioned after (2.4), there exists a sequence of
known constants
p{
{Cn,a}
(for any
Isn (0) I ~Cn,a Ie
0 <a <1),
such that
=O} ~a > {ISn (0) I >C n,a Ie =o}·)
(2.16)
1
where
n-'1: n,a -+A1"a /2 as n -+00 ,
1"S is the upper 100S% point of the standard normal df.
(2.17)
Let
then
eL,n =sup{b:
S (b) >C
}
n
n,a
vn
=nA(eu
,n
au ,n =in£{b:
Sn (b) < -C n,a }; (2.18)·
- eL,n ) /2Cn,a .
Then, it follows from Sen and Ghosh (1971) that
(2.19)
V
n
is a trans1ation-
invariant, robust and strongly consistent estimator of
v. Hence,
motivated by (2.14), we consider the following sequential procedure.
Let
nO be an initial sample size and
y(> 0)
constant (to be defined more precisely in Section 3).
be a positive
Define the
stopping variable by
(2.20)
Then,
"-
eN(c)
is the proposed sequential point estimator of
e
and
""
the risk for this estimator is
(2.21)
,
-6-
We are primarily concerned with the asymptotic behavior of
N(£),
@N(£)
and
in Section 3.
A*(£)
cztc l ~O)
(when
and these will be reported
In the remaining of this section, we consider certain
additional results on rank statistics and estimates, which yield
(2.12) and have a useful role in the derivations in Section 4.
Let
Xn, 1 S"'S Xn,n be the ordered rv's corresponding to
X1 "",X ' Then, it is known [viz., Sen (1959)] that for any positive
n
p « 00),
k
(2.22)
Elxl.IP <00 > Elx n,r I <00, V kip Sr Sn -kip +1 .
Also, if 0 <'1. slim infrln slim suprln S 1 - a l <1, then
n-+oo
n-+oo
ElxllP <00 -> lim supElx
Ik <00,
n-+oo
n,r
Further, note that the scores
a (i)
n
for every k ~O.
(2.23)
r
in (2.2) are nonnegative and
~=
nondecreasing (not all equal to 0) and hence,
<j>(u)du >0
(or
A >0).
This insures that there exists a sequence 0 {k } of positive
n
numbers, such that k =min{k sn} for which
n
(2.24)
L{i<k }an (i) >L{i>k }an (i)
n
n
1
and there exists an a: '2 <a <1, such that
n -1 k +a
as
n
[For
n
£
~0
(or 1),
(2.25)
{i< k }) is an empty set, so that
n
the corresponding sum in (2.24) is null. Also, note that fl <j>(u)du +0
as
k =n
n +00 .
{i >k } (or
and hence ~ >0
Theorem 2.Z
n
1-£
insures (2.25).]
Under (2.22), for every
k (> 0),
nO =nO(k) =min{n: n -k +1 ~k/p} where kn
n
'" ·Ik <00 for every n ~nO and
suah that E Ien
-e·
Then, we have the following
there exists an
is defined by (2.24) - (2.25),
lim supEle _elk <00, for every k ~O.
n-+oo n
In partiauZar, if p ~k, then Ee n exists for every
A
(2.26)
n
~1
.
-7-
PROOF:
By (2.4), (2.7) and (2.8), we have
[8(2) >x k ] <==> [S (X k) ~O], [~(l) >X k] <-> [S (X k) >0].
n
n'n
n n, n
n
n, n
n n'n
(2.27)
,..,
Let
for
1 be the rank of IXn,1. -Xn, kn I among Ix n, 1 -Xn, kn I,···, Ixn,n -Xn, kn I,
i =1 , ... , n. Then, by (2.3) and the monotonicity of the scores in
R.
(2.2) ,
Sn (Xn,k) = {-LU<k }an (Ri ) + LU>k }an (Ri ) }
n
n
S{LU>k }an(i) -L{i<k }an(i)} <0,
n
n
by (2.23). (2.28)
Hence, by (2.7), (2.8), (2.6), (2.27) and (2.28), for every
A
en
<X k
n, n
Similarly, for every n
n
~l,
with probability 1.
(2.29)
with probability 1.
(2.30)
~l,
en >Xn,n_k +1'
n
Then, (2.25) follows from (2.29), (2.30) and (2.22) -(2.25).
Q.E.D.
It follows from Theorem 2.1 that unlike the case of -X =n -lLn. IX.,
n
1= 1
k
n ~l,
the existence of Elx I does not necessarily require that
n
EIxll k <00; (2.22) and n adequately large suffice for
E I§n - elk <00.
We need to strengthen (2.25) to (2.12), under appropriate regularity
conditions.
Since
e
is not a martingale or reverse martingale, the
n
usual theorems are not applicable here.
We proceed to exploit the
asymptotic linearity (in b) of n -~2S (b)
for this purpose.
n
First, we note that by virtue of (3.4) of Sen and Ghosh (1973a),
for every
t
~o
and n
~ 1,
1
E{exp(tn-~Sn(O)) IH :
O
2
where An is defined by (2.6).
there exists an n (= nO(c,d),
O
e =O}
1
2 2
sexp{.2 t An}'
Hence, for every c >0
such that
(2.31)
and
d <0,
-8-
p{n-~ISn(O) I >c lognlHO: e =O} ~n-d, '</ n ~nO
Now, we denote by ep(r) (u) = (dr/dur)<j>(u),
assume that the
constants
r =0, 1, 2,
0 <u <1
and
ep(r) exists almost everywhere and there exist positive
K and
O
o«~),
such that
(2.32)
Further, we assume that the df F is symmetric, absolutely continuous
and possesses an absolutely continuous density function
first derivative
0,
flex)
for almost all
x
f(x) with a
(a.a.x), such that
(i) for
defined by (2.32),
sup f(x){F(x)[l _F(X)]}-o-n <00
for some
n >0,
(2.34 )
x
and
(ii) flex)
is continuous a.a.x. and
sup If' (x)
I <00
(2.35)
.
X
Define
B =BCep, F)
as in (2.10) and let
(2.36)
Then, following the lines of the proof of Theorem 2 of Sen (1980), we
arrive at the following:
Suppose that (2.33) -(2.34) holds for some
1" >0
and (2.34) holds.
and an integer
H :
O
nO
0 «4 +21")
Then, there exist positive numbers
(possibly dependent on
-1
1
« 4)'
d* and q
1"), such that under
e =0,
P{wn>n
-d*
(logn)IHO} ~ qn
-1-1"
,'</n~nO'
(2.37)
Now, by virtue of (2.3), (2.4) and (2.7) -(2.8), we have
!.:
~
A
P{nZ(en -e) >logn}=
. P{n- Sn ((logn)/v'ii)
~ole=o}
k
(lOgn) -S (0) +Bvnlogn] +n-Z[S
!.:
= P{n-z[S.
(0) -Bvnlogn]
n ~
n
n
~
1 .
!.:
1
P{Wn >2Blog n Ie =o} +p{n-ZSn(O) >2Blog n Ie =o}
>
.;:,... q*..n -1-1" , q*< 00, v.
v n - nO'
~ole
=O}
-9-
by (2.32) and (2.37), where we let
p{n~(en
bound holds for
d =1 +T,
- 8) < -log n}.
1
c ="2 B.
A similar
Then, we proceed to prove the
following
theorem 2.2
T
Under (2.22), (2.35) and (2.34) for some
for every
>0..
k 2
lim E{n /
PROOF:
,
k
k
-8I } = VkElzl ,
Ie
n
(2.40)
has the standard normal df.
Z
Let
I (A)
-8)\k=E{ln\e
1:A
2
+ E{ln (8
q >k,
A.
be the indicator function of the set
E{n~(en
For
-1
k < 2 (1 + T),
n~
where
o«4+2T)
n
k
-8)1 I(n
1:2 A
I8n
n
_8)lkl(n~len
Then,
-81:0; logn)}
-81> logn)} =J
(2.41)
n1
+J , say.
n2
by the Holder-inequality,
(2.42)
By Theorem 2.1, for every
E Ien - 8 Iq
<00,
for every
q >0,
there exists an
n ~nq'
n,
q
while by (2.39), for
such that
n
adequately
large,
n k/2 {pen1:2 I8 -8 I > logn)} 1-k/q
n
A
= 0(nk/2-(1+T) (l-k/q)) .
Since,
k <2(1 +T),
by choosing
q
adequately large, the right-hand
side of (2.43) can be made to converge to 0 as
J
n2
+
0
On the other hand, letting
as
n
(2.43)
n
+00.
Hence,
(2.44)
+ 00 •
1: A
Y =n 2(8 - 8),
n
n
we have (on letting
8 = 0)
J 1 =E{ IY ,kI ( IY 1 :0; logn)}
n
n
n
=E{ IY IkI( IY I :0; logn)I(w :o;n- d * (log n))}
n
n
n
+E{IY IkICly 1:0; logn)l(w >n-d*(log n))}
n
n
n
=J
+ J
,
nll
nl2
say,
(2.45)
,
'
-10whereby (2.37) and (2,45), for
J
n adequately large,
k
n12
S (log n) P{W >nn
d*
(log n)}
(2.46)
=O(n -l-T (log n) k) -+ 0
Further, for
IYn I S
as
n-+ oo
logn and w sn -d* (log n),
n
by (2.7), (2.8)
and (2.36),
BY =n-~ S (0) +R • IR I sn- d* (log n) .
n
n
n'
n
Finally, note that
1
In-~ (0)
n
1
I sn~An
1
=O(n~),
(2.47)
with probability 1,
_!.:
(2.31) holds and n 2Sn (0)/A is asymptotically N(O,l). Thus, by
some routine steps, it follows that under HO: 8 =0, for every fixed
k (~O),
logn)I(w Sn
n
-d*
k
k
(log n))}-+ A EIz I
.
(2.48)
Hence, from (2.45), (2.47) and (2.48), we have
lim J 1 =B-kzkElzl k = VkElzl k , V k ~O
n-+oo n
Thus, (2.40) follows from (2.41), (2.44) -(2.46) and (2.49).
Remarks.
~(u)
(2.49)
Q.E.D.
For the particular case of the Wilcoxon Scores (i.e.,
=U: 0 SU Sl), Inagaki (1974) has considered an almost sure (a.s.)
representation for
!.:
n 2(8 n -8)
A
interms of a sum of i.i.d.r.v's.
Our
Theorem 2.2 provides analogous results for a broad class of rank order
estimators.
Following the lines of the proof of Theorem 2.2, we have
A
-1
n(8 n -8) - B Sn (8) = ~,
say,
n
where under the hypothesis of Theorem 2.2, as n-+ oo,
(2.50)
(2.51)
Firther, it follows from Sen and Ghosh (1973a) that under
HO: 8 =0, {Sn(O), n ~l} is a (zero-mean) martingale sequence and under
conditions less restrictive than (2.33) -(2.34), there exists an n >0,
,
-11-
such that under HO'
A-lSn (0)
where
= Wen)
W={W(t), t £[0, oo)}
+ O(n~-n)
a.s., as
n
(2.52)
-+00,
is a standard Wiener process on
[0,
00).
From (2.50), (2.51) and (2.52), we conlcude that the Skorokhod-Strassen
{ V -1 n(8A
embedding of Wiener proeess holds for
hypothesis of Theorem 2.2.
n
-8)
}
under the
Sen and Ghosh (1973b) have also condidered
an a.s. representation for one-sample rank order statistics.
special case of their theorem, we have under
Sn (0) = I~1= l</>* (F (X.1 )) + ~*,
n
As a
HO: 8 =0,
n ~ 1,
where, under conditions less restrictive than the ones in Theorem 2.2,
n-~~; =O(n-n) a.s.,
as
n
-+00
(for some n >0).
(2.54 )
From (2.50), (2.51), (2.53) and (2.54), we conclude that
n-~In(~n -8) -B-II~=1</>*(F8(Xi))X-+() a.s., as n -+00,
(2.55)
and this extends Inagaki's theorem to a broad class of signed rank
statistics where
</>*
need not be bounded.
3•
THE
MAIN TIIEOREMS
A
The asymptotic behavior of N(£), 8N(£) and A*(£) (as c 2/c l +0)
will be studied in this section. The results to follow depend on
through
c =c 2/ c l
only, and hence, for notational simplicity,
A
we let
0
A
c l =1, c 2 =c, N(£) =Nc ' 8N(c) =8 ' A*(~) =A~,
Nc
o
An ()
=Ac • Then, we have the following.
c (c)
~
o "'"
Theorem D.l Under (2.22) and (2.33) -(2.34) for some
no (£) =nc
~
y>O
[in (2.20)], as
cS
and
1
<'4' for every
c +0
(3.1)
(3.2)
.
·.
-12-
Theopem'3.2 Undep (2.22), (2.35) and (2.33)
os(4+21")
-1
-(2.34)~
,1">1; 1+2y<1",
when
y>O,
(3.3)
whepe y is defined in (2.20), then
lim (A*/Ao ) = 1 •
c+O
c c
(3.4)
It may be remarked that (3.4) asserts that the pisk involved
in the proposed sequential procedure is asymptotically (as
c + 0)
equal to the risk of the corresponding optimaZ fixed-sample size
procedure.
Thus, for all
F and
satisfying the hypothesis of
C/>
Theorem 3.2, the proposed sequential procedure is asymptoticaZly pisk-
f11 C/>(U) Irdu <00
efficient.
In particular, (3.3) demands that
some
and this is true for the Wilcoxon, normal scores and the
for
°
r >6
r
other commonly used rank statistics and is less restrictive than
exp{t(j>(u) }du <00
(for some
t >0), as employed in Sen and Ghosh
o
(1971) for the confidence interval problem.
The asymptotic normality of
(no)-~(N
_no)
c
c
c
depends on the
asymptotic normality and uniform continuity, in probability, of
~
{n 2 (V -v)}.
n
~
.
For Wilcoxon scores, the asymptotic normality of n 2(V -v)
n
has been studied by Jure~kova (1973) and her treatment holds generally
for bounded and continuously differentiable score functions; however,
for unbounded scores, this remains as a challenging open problem.
The following theorem presents the impact of this on the asymptotic
normality of stopping times.
Theopem
3.3
Suppose that in (2.20)
Y
loP
>-2' Nc /n c ~ 1 as
c + 0 and
n ~2(vn -v)/f3-:;r-N(0, 1),
whepe
{ n2lvm-vnl
~
} ~Oas
p
sup
0+0,
m: Im-nlson
(3.5)
13 is a finite positive numbep. Then, as c +0,
( n0) -~ (N -n0)
c
c
c
---rr N( 0,
f32/ v 2) •
(3.6)
..
-13Note that (3.5) may be replaced by the weak convergence of the
1
{n-~(vk -v)/13: k sn}
partial sequence
to a Gaussian function.
The
proofs of the theorems are presented in the next section.
4.
We let
n
o
c
PROOFS OF THEOREMS 3.1, 3.2 and 3.3
= [c
-!.:
(see (2.15)), and, for every
2V ]
0 <E <1,
we
let
n
lc
=[c
-1/2(1+y)
where, we choose
c
]
n
'
0
so small that
definition in (2.20),
and
=[(l-E)n]
c
2c
nO sn
lc
n
<n
o
3c
2c
(4.1)
= [(1 + E)n ],
c
<n 3c '
Then, by
with probability 1, so that
Nc <::n l c '
(4.2)
as, for
n
n
~
2C
~O
.!.:o
' C"ll -v sc~C (1 -E) -C"llO"'" -EV.
Now, by (2.16) -(2.19)
and proceeding as in (2.38), it follows that, for
~ ~
{
Pn(t1
~ A
{
Pn
(e
j .
u,n
L,n
}
-e»logn=O(n
-l-T
n
adequately large,
(4.3)
).
-e) <-logn } = O(n -l-T ),
(4.4)
so that by (2.17), (2.19), (4.3), (4.4), (2.36) and (2.37), we
conclude that for
n
p{ Iv
n
adequately large,
l
;<:EV} s3q*n- - T ,
-vi
Hence, by (4.2) and (4.5), as
P{Nc sn 2c } s
\L
n
lc
{3q*n -l-T}• = 0 (n -T )
snsn
P{N
c
>n} =P{k
n
(4.5)
c +0,
1c
2c
= 0 (c T / 2 (1+ y )) ,
In a similar manner, for
q* <00.
b Y (4 • 1) •
(4.6)
~n3c'
<c-~(Vk +k- Y),
1
~P{n <c-~(V
V k E [nO' n]}
1
+n- Y)} =p{vn >c1l-n- Y} .
n
!.:
!.: 0
Y
= p{V - v > c ~ - c ~- n - }
n
c
sp{
IVn -vi
>n}
where by (4.1),
n >
(4.7)
o.
,
"
r ,
•
-14Thus, by (4.5) and (4.7),
P{Nc >n3c } + 0 as c +0. Since E(> 0) is
o
arbitrary, by (4.6) and the above, p{ IN /n -11 >d + 0 as c +0.
c c
Also, by (4.5) and (4.7), I >
peN >n) = 0(n3-T) +0 as c +0, and
n-n 3
c
o
c
0
hence, E{Nc I(Nc ~n3 c }}/nc +0 as c +0. Also, E{Nc I(Nc ~n2 c )}/nc ~
(1 -E)P{N
~n2
by (4.6), while by (4.5), (4.6) and (4.7),
E{N I(n
)}/no can be made arbitrarily close to 1 by choosing E
c
2c
c
3C
small. Hence, E(N /no ) +1 as c +0. A similar proof holds for
c
c
<n <n
} +0
c
c
E(N /no)k + 1, V 0 ~k <1.
c
c
This completes the proof of (3.1).
Now,
it follows from Sen and Ghosh (1973a) that under
H : 6 =0, {Sn(O)} is
O
a martingale sequence, and hence, using the Kolmogorov inequality (for
submartingales) and some standard analysis, we obtain that as
lim{
max
n-~Is (0) -S (O)I} =0,
0+0 m: Im-n I~on
m
n
in probability.
By (2.39), (2.47) and (4.8), we obtain that as
lim{ max
n~ 11~ - I} = 0,
0+0 m: Im-nl~on
m n
e
n too,
n+ oo ,
in probability,
so that (2.9), (4.9) and (3.1) insure (3.2).
(4.8)
(4.9)
Hence the proof of
Theorem 3.1 is complete.
To prove (3.4), we make use of (2.15), (2.21) and (3.1) (i.e.,
limE(N /no ) =1), and hence, it suffices to show that
c+O
c c
k
1
A
lim(vc 2) - E(6 -6)
N
c+O
c
2
= 1
(4.10)
Now, by Theorem 2.2, for every k E (2, 2 + 2T),
n
E{(§N -6)2I(N ~n2 )} =I 2c E{(e _6)2 I (N =n)}
'c
c
c
n=nlc
n
c
n
~In:~ (EI~n _6I k )2/k(p{N =n})1-2/k
c
lc
n
.
e _6 Ik) 2/k (P{Nc ~n 2c }) l-2/k
2c E I
Ln=nlc
n
:c;; [\'
=(0(ni~k-2)/2))2/k(0(cT/2(1+y)))1-2/k
=0(c(k-2)(1+T)/2k(1+y)).
[by (4.1)]
(4.11)
[by (4.6)]
f
-15Since, by (3.3),
~ >0,
(1 +T) >2(1 +Y),
while for
T =1
1
2~ - n
1
(k -2)/k ='2 + 8 +4~ -2n > '2 for
n >0,
(4.11) that by a proper choice of
+~
n > 2~
k £(2,2(1 +T)),
and
k =2(1 +T) -n,
we obtain from
under (3.3),
k 1
I'
2
lim(vc 2) - E{ (eN - e) I(N :;;n )} =0.
2c
c+O
c
c
Similar1y,as
E{(e
Nc
(4.12)
c+O,
-e)2 l (N
~n3 c )}:;; (L>
n-n
c
Ele
3c
k
_el k12 / (P(N
J
n
c
~n3 c )) 1-2/k
=0(c(2+T)(k-2)/2k),
so that noting that by (3.3),
T >1
(4.13)
and then taking
k =4,
we
obtain from (4.13) that
k
lim(vc 2)
c+O
-1
A
E{ (eN
2
- e) I(.N_ ~n3 )} = O.
c
c
c
(4.14)
Thus, to prove (4.10), it suffices to show that
k
lim(vc 2)
c+O
-1
A
E{ (eN
c
2
- e) I(n
<N
2c
<n
c
3c
)} = 1.
Now, by Theorem 2.2 and the definition of
lim (vc
k -1
2)
dO
while,
P{n
2c
<N
c
<n
3c
} + 1
as
c +0.
nO,
A
{
c
(4.15)
we have
4 k
E( S
- e) } 2 < 00 ,
nO
c
(4.16)
Hence, to prove (4.15), it is
enough to show that
k -1
A
A
2
lim(vc 2) E{[(SN -S ) l(n
<N <n )} =0.
2c . c
3c
c+O
c
n0
(4.17)
c
Let us then write (as in the proof of Theorem 2.2)
k
A
Bn 2(S
so that writing
:;; ER 2 l(n ~A
Ie -e I
n
n
_k
- S) = n 2S (S) + R ,
n
n
n
(4.18)
2
2
k A
2
k A
ER =ER I(n2le -el:;; 10gn) +ER l(n 2 lS -si > 10gn)
n
n
n
n
n
2
2
kA
I
:;; logn) +2B E{n(S -e) l(n 21s -s > 10gn)} +
n
n
1 2 k A -e I. > 10gn)}
2E{n -Sn(e)l(n2ISn
through (2.47) (where we take
A
and then proceeding as in (2.42)
k =2
and
T >1 + 2y, Y >0), we obtain
"
J
.
-
., "
-16-
by some standard steps that
ER 2
n
= 0(n- 1 - y )
(4.19)
'
so that
(4.20)
Further, for
-1
N
n 2c <Nc <n 3 c '
C
:5:n
-1
2C
k
=O(c 2).
Hence, by virtue of
(4.18) and (4.20), it suffices to show that
lim{(nO)-lE[(SN (0) -S (0»2I(n
<N <n )
3c
2c
c
c
c +0
c
n0
c
Since, under
H :
O
e =0,
(3.1),
<N
<n
P{n
2c
c
3c
}
{Sn(O)}
1
-r
as
(no)-l\,n~c
c
Ln-n
:5:
n
c+o
and (2.31) holds, we have
Ie
=0]
c
(E[(S (0) -S
n
2c
=O]} =0. (4.21)
is a (zero-mean) martingale, by
O
(n )-lE[(SN (0) -S (0»2I(n
c
oc
z <N c <n 3c )
c
Ie
n0
c
(0»4 Ie =O])\P{N
3C
:5: (nO)-1(\,n
E[(S (0) -S (0»4 Ie
c
Ln=n 2C
n
nO
c
=n})~
=0]J~(p(n2 c:5:Nc
:5:n
3c
»~
c
(n~)-1[~:~2CO(ln -n~I»)~-l
:5:
o
= 0((n3c -n 2c )/n c ) ,
where by (4.11),
small by choosing
(n
3c
-n
2c
E so.
(4.22)
)/n
o
c
~
2E
and this can be made arbitrarily
This proves (4.21) and the proof of Theorem
3.2 is complete.
To prove Theorem 3.3, we note that by definition in (2.20),
c-~N
c
:5:N :5:C-\V -1 + (N -l)-Y)
c
N
c
c
whenever
N >n '
O
c
(4.23)
1
Thus, noting that
n~ . . . vc-~
we have from (4.23), for
N >n '
c
O
f
-17-
as
c
to.
Hence, (3.6) follows from (3.5) and (4.24).
5.
Q.E.D.
SOME CONCLUDING REMARKS
Under the hypothesis of Theorem 3,2, the sequential procedure
is asYmptotically risk-efficient.
It may be noted that
depends on the score function through
V
=A/ B,
by (2.6) and (2,10) and are functions of
$
where
A and B are defined
(and F).
To make this
depence clear, we denote
(5.1)
where
defini~g
A$2 =
$*
as in before (2.2),
11 <p 2 (w)du
and
B$ =
o
f°fx
(5.2)
$* (F (x)dF (x) .
_00
Thus, if we have two different score functions say,
the corresponding optimal
0
11.
A(c l , c 2 , A$l' B$l)
c
$1 and $2'
then
are
and
A(c I , c 2 , A$2' B$2) ,
(5.3)
and smaller is the quantity, the better is the corresponding procedure.
Hence, the relative efficiency of the procedure based on the score
function
$2
with respect to the one based on
$1
is
e($l' $2) =A(c l , c 2 ' A$l' B$l)/A(c l , c z ' A$2' B$z)
= (A$IB$2)/(A$2 B$1)
(5.4)
and this agrees with the (square root of the) classical Pitmanefficiency of the rank tests (for location) based on the score
function
$2
with respect to the score function
define 1jJ(u) =1jJ* ((1 +u)/2),
0 <u <1,
A~
= f(ff/f)2 dF
Hence, if we
where
1jJ*(u) =_f(F-I(U))/f(F-I(u)),
then
$1'
0 <u <1,
is the Fisher information and by (5.4),
~,
"
.c- ..
(tl
'
-18e(~, ~)
= p(~,
=
~nd
~)
(5.6)
(5.6)
(to~*(UN*(U)dU)/Al~)
the equality sign holds iff
~* .e~*
•
Thus,
~
is an optimal score function.
For the procedure based on the sample means and variances,
considered by Starr (1966) and Ghosh and Makhopadhyay (1979), the
corresponding
is
(5.7)
Thus, the asymptotic relative efficiency of the proposed sequential
procedure with respect to the normal theory procedure is
e(~, N) = aB~/A~
(5.8)
.
In particular, if we use normal scores (i.e.,
the standard normal df), then, (5,8) is
equality sign holds only when
~
1
F is normal.
~*(u),
for all
the inverse of
F,
where the
This expa1ins the asymptotic
supremacy of the normal scores procedure over the parametric procedure.
Even for Wilcoxon scores, when
and it is usually >1
F is normal, (5.8) reduces to
for distributions with heavy tails.
1~
(3/n)'2
For both
these scores, conditions for the applicability of Theorems 3.1 and 3.2
hold, while Jure~kova's (1973) theorem insure (3.5) for the Wilcoxon
scores.
~
.978
t
-19REFERENCES
[1]
CHOW 1 Y.S. and ROBBINS 1 H. (1965). On the asymptotic theory
of fixed-width sequential confidence intervals for
the mean. Ann. Math. Statist. ~R1 457-462.
[2]
GHOSH, M. and MUKHOPADHYAY 1 N. (1979). Sequential point
estimation of t-e mean when the distribution is unspecified.
Commun. Statist. SeT' A. ~, 637-652.
[3]
INAGAKI,
[4]
JURECKOVA, J. (1973). Central limit theorem for Wilcoxon rank
statistics process. Ann. Statist. 1, 1046-1060.
[5]
ROBBINS, H. (1959). Sequential estimation of the mean of a
normal populat:i,on. PT'obabiUty and Statistics
(H. Cramer Volume), Almquist and Wiksell, Uppsala 1
pp. 235-245.
[6]
SEN, P.K.
(1959). On the moments of the sample quantiles.
Calcutta Statist. Assoc. Bull. ~. 1-19.
-[7]
SEN, P.K.
(1980). On almost sure linearity theorems for signed
rank order statistics. Ann. Statist. ~, in press.
N.(1974). The asymptotic representation of the
Hodges-Lehmann estimator based on Wilcoxon two-sample
statistic. Ann. Inst. Statist. Math. £2, 457-466.
v
.;
[8]
SEN 1 P.K. and GHOSH, M. (1971). On bounded length sequential
confidence intervals based on one-sample rank order
statistics. Ann. Math. Statist. i£, 189-203.
[9]
SEN, P.K. and GHOSH, M. (1973a). A law of iterated logarithm for
one-sample rank order statistics and an application.
Ann. Statist. 1, 568-576.
[iO]
SEN, P.K. and GHOSH, M. (1973b). A Chernoff-Savage representation
of rank order statistics for stationary $-mixing
processes. Sankhya3 SeT'. A. ~~, 153-172.
[11]
STARR, N.
[12]
STARR, N. and WOODROOFE, M. (1969). Remakrs on sequential point
estimation. PT'oc. Nat. Acad. Sci. USA 2~, 285-288.
[13]
WOODROOFE, M. (1977). Second order approximation for sequential
point and interval estimation. Ann. Statist. ~, 984-995.
(1966). On the asymptotic efficiency ofa sequential
procedure for estimating the mean. Ann. Math. Statist.
~I, 1173-1185.
lit.'
,..