Sen, P.K.; (1979).The Cox Regression Model, Invariance Principles for Some Induced Quantile Processes and Some Repeated Significance Tests."

•
•
THE COX REGRESSION MODEL, INVARIANCE PRINCIPLES FOR SOME INDUCED
QUANTILE PROCESSES AND SOME REPEATED SIGNIFICANCE TESTS
by
PRANAB KUMAR SEN
Department of Biostatistics
University of North Carolina at Chapel Hill
Institute of Statistics Mimeo Series No. 1208
.
,.
r
January 1979
THE COX REGRESSION MODEL, INVARIANCE PRINCIPLES FOR SOME INDUCED
QUANTILE PROCESSES AND SOME REPEATED SIGNIFICANCE TESTS*
PRANAB KUMAR SEN
University of North Carolina, Chapel Hill, NC
•
ABSTRACT
For the Cox regression model, the partial likelihood functions
involve linear combinations of induced order statistics.
Some in-
variance principles pertaining to such linear combinations of induced
order statistics are studied and the theory is incorporated in the
formulation of suitable repeated significance tests (for the hypothesis
-e
of no regression) based on these partial likelihoods.
•
properties of these tests are also studied .
AMS Subject Classification Nos:
Key Words and Phrases:
Asymptotic power
60B10, 60FOS, 62G99.
Analysis of covariance, asymptotic power,
contiguity of probability measures, Cox regression model, hazard rate,
induced order statistics, quantile processes, repeated significance
tests, survival functions, Wiener process.
* Work supported by the National Heart, Lung and Blood Institute,
Contract-NIH-NHLBI ...71-2243 from the National Institutes of Health.
2
1.
INTRODUCTION
I
14
In the Cox (1972) regression model for survival data, it is
assumed that the i.th subject (having sur1)i1JaZ arne
covariates
(given
Z.
~1
for some
and a set of
has the hazaY'd rate
p? 1)
=z.)
~1
(1.1)
hi (t)
where
Y.1
ho(t),
=hO(t) exp (~' .ti) ,
the hazard rate for
z.
~1
i
=0,
=1 ~
.•• , n,
is an unknown, arbitrary
~
B ) parameterizes the
p
regression of survival time on the covariates. We assume that
the
Yi
t
'
~'= ((31' " ' ,
non-negative function and
is continuous in
.
almost everywhere
hO(t)
(a.e.), so that ties among
may be neglected, with probability 1.
Let
~ = (QI' ••. ,~)',
the vector of anti-ranks, be defined by
Y
(1. 2)
Q1.
where
Y
nl
< ... <Y
nn
=Yn1.,
1:::; i :::; n ,
are the order statistics corresponding to
follow-up (censoring), the partial (log-) likelihood function when all
the failures have been observed [c.f. Cox (1972, 1975)] is given by
logLn =L:=l{~'~i -lOg[Lj=iexp{J!'~j}J} .
(1.3)
To incorporate censoring, Cox (1972) considered a more general
case where there are
m(~
1) ordered failures
(a subset of
t , ..• ,t
l
m
such that at time t. - 0, there is a risk set R. of r.
n1.L
J
J
J
subjects who have neither failed nor been censored by that time, for
the
Y
i :::;j
~m
(so that
times for the
(1,4)
n
t. =Y *
Qj
J
.
'
YI , ... , Yn · Then, following Bhattacharya (1974), ~l' ..• , ~n are
termed the induced order statistics. In the event of no loss in the
R c ... c R )
mI'
If
WI' "', Wn be the censoring
subjects, then, we have
and
R. = {i:
J
Y~ =Y. AW. ~t.,
1
1
1
J
l:::;i :::;n},
l:::;j :::;m,
3
Q* = (Qi~ ••••
where
~1'
is asub-vector of
a Ab =min(a, b),
the COX (J972) partial (log-) likelihood function is
Then~
(1. S)
logL*nm
=L~=l{~'~*
•
J - ' < j -log(L11.'R i exp{~'~.})}
1
Yi =Vi>
[Note that if m =nand
•
and
Q
l<i':n,
then
L*
nm
=L.
n
A
d'i[Jere/;c~
time version of (1.5) has also been considered by Cox (1972) and we
shall refer to that later on.]
For testing the hypothesis of no regression viz ••
(1.6)
8
~
=0 vs.
~
Cox (1972) considered the test statistic
L*
(1.7)
nm
= U*, J* - U*
""l1.m ""11m •
"'11JII
where
(1.8)
Unm
*
= (a;:y/ .......,
aS)logLnm
*
I
Q
IJ=
~
and~-
•
0'
~
*
J""11m
*
= (a~2/ar<ar<')logL
~~
nm
stands for a generalized inverse of
~.
I
R 0
f;:,=~
•
Cox (1972; 1975,
Section 5) argued heuristically that under
HO' L*nm has asymptotically
chi-square distribution with p degrees of freedom (DF).
In a variety of situations, relating to clinical trials and life-
testing experimentations. one may be interested in monitoring the
study from the beginning with the objective of an early termination if
H in (1.6) is not tenable. Such a plan is known as a progressively
O
censo~ed scheme (PCS) [viz .• Chatterjee and Sen (1973) and Sen (1976.
1979)].
Thus, in a PCS, instead of making a terminal test at the m-th
failure
t
,..
a
t.
J
(j
~
1)
J
m
one may like to review the process at each failure
and stop experimentation as soon as
(defined as in
nj
leads to the reJection of
L*
and J * )
nj
for some j ~ m; if L* 1 • " ' , L*
n
nm are all insignificant, then
is accepted along with the termination of the study at the preplanned
(1.7) - (1.8). but., based on U*.
~nJ
4
m-th failure.
Hence~
a repeated significance testing (RST) procedure
Since the
is involved in a PCS.
L. are neither independent nor
nJ
have independent increments, more elaborate analysis is needed to
find out the critical values having prescribed overall level of
significance,
•
We may note that by (1,5) and (1.8),
(l.9)
U*k
n
=t-l{~*
-r~lIR.Z.}
J~j
J 1£ J~l '
for
k =1, •.. ,m ,
and hence, these are all linear combinations of induced order
statistics.
For this reason, we first study some invariance principles
relating to induced order statistics and then incorporate these in
the study of the asymptotic properties of the proposed RST,
These
induced quantile processes are introduced in Section 2 and their weak
convergence results are presented there.
Parallel results for the
discrete time case are treated in Section 3.
Section 4 is devoted to
the study of these invariance principles under (local) contiguous
alternatives.
The concluding section incorporates all these results
in the study of the asymptotic properties of the proposed RST procedures
relating to the Cox model.
2.
WEAK CONVERGENCE OF SOME INDUCED QUANTILE PROCESSES
By analogy to (l,3), (1.5) and (1.8), we let for every
(2.1)
(2,2)
logLnk =
L~=l{~'~.J -IOg{L~=jexp{~'~.})},
1
I
!lnk = (~/1~) logLnk ~=Q
=L.Ilkj =1{ ~.
.
- (n - J + 1)
J
(2.3)
:!nk = - c~2/]~' )logLnk
=L~=l(n
-j +1)-ICn
-l\'n
L. i =j
~.
}
1
I~=O
-j)~j'
say,
k:
I
~k ~n,
.
I
5
s
where
"'1ln
and
=0
S . =(n _j)-lI~=.(~ -Z-~)(~ -f'.l')"
(2.4)
"1lJ
1
J
.I
'<j
.
~
-l\n
~j=(n~J+l)
(2.5)
1 -:::j <n -1;
'<:i
l.i=j~.,
l$j~n.
1
Conventionally, we let
•
uo=o,
"'1l
(2,6)
Vn~l.
JO=O,
"'1l
~
~
First, we study suitable invariance principles relating to the triangular
arrays
0 ~k ~n;
{!!nk:
n
~ l}
and
0 ~k ~n; n ~ l},
{:!.ok:
~ll
Throughout this paper, the covariates
... ,~
are assumed to
be stochastic vectors; there are some simplifications when these are
nonstochastic and these will be briefly considered later on.
Also, in
the usual custom of an analysis of covariance model, we assume that
~l'
•..
,~
are independent and identically distributed random vectors
(i.i.d.r.v.) with
l!. = E~, [= E(~ -l.9
(2.7)
(~
-l!.)'
(both exist),
and
(2.8)
is positive definite (p.d.) with det. [<00.
[
S 1 = (n - 1) -1\. 1 (Z. - Z HZ. - Z ) ,
For clarification of ideas, we note that
(where -Z =n -ll:n. lZ')'
"'11
1= ~
while for
2
~
J'
l.1
"1l
~
=
~l
~l
"'1l
n - 1,
l
.)-l{\n
Z Z' S - (
l.' 1 Z• Z'. - dL.' l;::.n;::.n
"'1lJ. - n - J
1= ~1~1
1= ~i ~i
(2.9)
Z d- 1 7 J(\n
Z' d- l Z')} '
(n - J. + l)-l(\n
l.i=l~i - l.i=l~.
Li=l~i - l.i-l~.
1
and
S
"1ln
= O.
Thus, if
Zl' ••. , "1l
Z
~
of the experimentation, then
&
~k
are all observable at the beginning
depends only on
hence, is observable at the
(k -l)th
~k
and
depends on
Qk+l' ... ,~).
~l'
.•• ,
~
1
failure, for
Ql' ••. ,Qk
Qi' ..• ,Qk-l
2
~k ~n,
and
Also,
(put not individually on
We are primarily interested in the asymptotic behavior
"'1l
6
of the two partial sequences
Let
0::; k ::; n}
Zl"'"
~
.
for
Z
~T1
is nondecreasing in
B
nk
introduce a sequence
{:lnk ~
and
be the
•.• , Z ,
"11
and
Then
{llnk'
l<,k<n,
t (E
a-field generated
while
Hn 0
=B(Zl'
~,
•.• , Z ).
~11
.
For every n (~ 1) 1 we
k (<::; n) •
{k (t) ~
n
0::; k ::; n} ,
= [0,
lJ}
of nondecreasing and
right-continuous non-negative integers by letting
(2.10)
l J k)::; pt}, t (E.
Trace (J"'11n .....n
kn (t) =max{k:
Let then
r
~n
= {t:~n (t),
t (E}
be defined by
(2.11)
Note that in (2.10) and (2.11), we have assumed that
~
J- 2 = B
""nn "'Il
is defined by
B J B'
"11"11n"11
=~
I
,
.J
"11n
the identify matrix of order
Later on, we shall see that by virture of (2.8),
J
"11n
probability, and it is also possible to replace
k (t)
(where
denotes the largest integer
[sJ
n -~ -~!![ntJ'
r
t
£
E.
is p.d. and
is p.d., in
n
by
t: (t)
::; s) and
p.
~n
[ntJ
by
Then, the main theorem of this section is the
following
Under (2.7), (2.8) and
Theorem 2.Z,
to
and
$.
={$.(t) ,
t
£
E}
t: (j) = {~( j) (t),
f (t ) = (t: ( 1) (t),
where
t
£
j =1, ..• ,p
E},
standard Wiener process on
~ = Q,
H :
O
~ weakl:y.oonveztgea
... , ~ (p) (t)) "
t
£
E
are independent copies of a
E.
Before we present the proof of the theorem, we consider the
following lemmas.
Lerrorza 2.2.
Under C2,7) and HO :
martingaZe (for every
n ~ 1) •
~
=Q,
{!!ok' B
nk
,
O::;k ::;n}
is a
,
,
7
Note that by
Proof:
B _ ,
nk 1
(n-k+l) -1 ,
V 1
Ho'
under
(1, .•• ,n)\(Ql' .•• ,Qk.,.J)
E(~
so that
+l)-lL~=k~..
(n -k
k
~
1t
-1,n
~
Now, given
•
for every
U k ~,U k 1 = ~ - Cn - k + 1)
~ -,
~k
(2.12)
set
(~, 2),
~k
Qk
l.' -k~
~-
•
~!
can take on anyone value in the
with the equal conditional probability
IBkl)=(n-k+l) -l{,n
,k-l} =
L'-lZ'-L'-l~
n 11
1~i
Thus, under
J
HO'
E(l!nk -l!nk_lIBnk_l) =0 a.e.,
Q.E.D.
~k ~n.
H '
O
k
E{!lnk-l!nk-1)(l!nk-!lnk-l)'IBnk - 1 } = nn_i +1 ~k' V
By the same arguments, it follows that under
(2.13)
~k
where the
are defined by (2.4).
!1 (~, J?)
(2.14)
=
l~k~n,
Note that if we let
}(~ - J?) (~ - J?) ,, V~, £ £ RP,
then we have
(2.15)
S
"1lk
= (n
,
<P(Z, Z )
k<'
~.
<
_
-l-J_n ~.1 ~.'
J
- k + 1))
2
1 ~k ~n - 1 •
V
As such, by the same arguments as in the proof of Lemma 2.2, we have
H :
O
under
8 = ,..,0,
,..,
(2.16)
= S ,
E(S
"1lklBnq )
~q
Also, note that
S
~n1
Lerruna 2. J.
ES
"1l 1 =r.
,..,
Under (2.7) and
{Sk"1l
,..,r , Bn k ; l~k::;;n-l}
For a
H :
O
~
=Q,
Let
{k}
n
(2,18)
{11~k .... 1:
II,
but
B ; 1 ~ k ::;; n - 1}
nk
n
which
is a martingale.
-J 2
k -+0
n
as
II~II =max{laijl:
n(~ 2), under
l"i,
H : ~ =Q '
O
a non-negative submartinga1e.
be any sequence of positive numbers
k -+00
n
~
n (~ 2) ,
for every
~=((aij))' we let
pxp matrix
,..,1
This leads us to the following
Then, by Lemma 2.3, we have for every
(2.17)
2 , .•. , 2 ,
is aU-statistic (based on
are LLd.r.v.), so that
.
l~q~k~n'-l
'V
(k
n
~n),
such that
n -+00 •
Then, by (2.17), (2.18) and the generalized Ko1mogorov-inequality for
submartinga1es, under
H :
O
~
= Q and (2.7),
j "p},
8
(2.19)
(2.20)
~
Note that under
(1, ... ,n)
every
k:
~ =
H :
O
E -IE
fl,
II S
k
"7ln-
(Ql"
~ r II ~
n
•. ,~)
~n-k =g(~
E
•
>0 •
takes on each permutation of
(n!)~l. Also, by (2.15), for
with the common probability
1 ~k ~n -1,
V
~
, ... ~ ZQ )
Tl .. k
is aU-statistic
n
[Hoeffding (1948)], so that
(2.21)
E II ~m _k - I" = E II g (~
~ ... , ~ ) - III
n-k
n
=E{E[llg(~
, ... , ~ ) -III IBnOD
n-k
n
=E{(k;lfl.
l~ll<"
I.
'<lk+l~n
11~(~i'1 ""~ik+l ) -rll}
=Ellg(~l' ''''~k+l) -III.
Thus,
EII~n_l -III =EII1!(~l' ~2) -III
•
by (2.7) and
<00
(2.22)
On the other hand,
m ~ 2}
{U (Zl' ... , "1ll
Z ),
~~
is a reverse martingale
sequence, so that by the reverse sub-martingale convergence theorem,
Ilg(~l' ... ,~) - III
converges in the first mean to
Hence, (2.22) converges to
0
as
righthand side of (2.19) is
n
-+00,
(7,23)
l~k~n
II.!. J
=p{ max
l~k~
~p{
max
n "7l
k -
(or
k)
n
-+00.
Consequently, the
0
as
E > 0,
~ r II
n
>d
IIIn £1=
\~ l{(n -i)(n -i +l)-l(S
.
.
"7l1
l~k~n-l
m -+00.
and of (2.20) converges to
Note that, for every
p{ max
n
0 as
-0
+ (n -i +l)-lnll
>d
~
,~
1
Iln-lt=l(S k -I) II >E/2} +p{IIIII >2 En},
1
"7l
9
where the second term on the righthand side of (2.23) converges
to
0
On the other hand, using the inequality that
(2.24)
Iln-lI~_l (5
max
l~k~n-1
..
~
•
. -
1--TI1
D II
k
II~k - r: II} + :
{max
l~k~n-k
max
115 k n-k <k~n-l ~
{
n
r: II}
n
we obtain on using (2.18), (2.19). and the convergence results following
(2.22) that the first term on the righthand side of (2.23) converges to
o as n ....
00.
Lemma 2.4.
This leads us to the following.
Under (2.7) and H :
O
(2.25)
we
max
l~k~n
II n- ~ (J
~
~ = Q.
k - k r)
,.."
II L
0, as n .... 00.
By (2.8) and (2,25),\\Te conclude that under HO: ~=Q,
1
(2.26)
n- :lnn L
and is p.d.~ in probability.
r
Note that by (2.2), with probability I,
max Ilu k -U k-1 11 = max II~
-TI
l<k<
~k
_ _n -TI
_ _n
l <k<
(2.27)
::; 2{ max II~.II} =2Z*,
n
_l_n . 1
1 <'<
where, by (2.7). for every
say
E: >0,
p{Z* >
n
(2.28)
,. (n -k + l)-IL~=l~
II
1 ~.1
E:Tn'l .... 0. as
n .... oo
•
[The proof of (2.28) follows from well known results on the maximum
of n
LLd.r.v. 's and hence is omitted.]
lin
Then, for every
E: >0,
-1 n
I E{(U-TIl. -u-TI1.. I)(U-TIl. -u-TIl. 1)'1(lI u
. 111 >E:rn)IBn1.. III
i=1
-TIl. -u~1-
1 n
=lln- . 1 n
1=
I
_~1.
n
+
1
L (~Z . -1~)(~Z
-f~)'I(II~Z
-1~11
1
j
1
j
1.
.•
J=1.
J
.
::; (n -1 ~l. Trace
[Cn - i) (n - i
.1= 1
+ 1)
1 r
5.] ) 1CZ * >-2
E:rn)
NTIl
n
'
-1
1 E:rn)
r
= (Trace [n -1 J-TIn ]) 1(Z n* >-2
...£. 0, as n .... oo [by (2,6) and (2,28)].
>E:rn)11
.
I
10
Let us now return to the proof of Theorem 2,1,
By virtue of
(2,10), (2,11), (2.13), (2,25), and Lemma 2,2, for every
0
<t·, I,
<:5
(2.30)
I
-~
=J
"'11n
-~
=:lnn
{k
n (s)
I I =l
.
} -"2
1
n - 1.
.'. J
n-
1 +
-~
:lnkn (s) :lnn
1
~I
"'11n
•
.E.~ s1 = (s At)1
7
p
p
=E{s(t) [s(s)] I} ,
"'"
"'"
This, intuitively suggests the process
f
in Theorem 2.1.
To prove
the desired weak convergence result, we make use of the martingale
property in Lemma 2.2 and appeal to functional central limit theorems
for (a triangular array of)
martingales [viz., Scott (1973)]; we only
need to extend these to the multivariate case, but that poses no
problem [the usual Cramer-Wold type arguments filter through without
demanding any extra condition].
Our Lemma 2.4 and (2,29) insure that
both the conditions of Scott (1973) hold in this multivariate case and
the proof of the theorem is
the~fore
complete.
and (2.25), it follows that we may also take
sn (t)
1
By (2.10), (2.11),
k (t) = [nt] ,
n
t
£
E
and
1
=n-~r-~[ t]'
"'"
"'" n
t £E.
Note that by virtue of the Courant theorem (on the ratio of two
quadratic forms) and by (2.25),
(2.31)
max
-1
U k -1 I
IU'kJ -1 U k/u'k(nr)
"'11 "'" "'11
1 s:k ~ "'11 "'11n"'11
l
s:max{IChl(nrJ- )
""""11n
p 0 " .as
~
(where
roots) ,
chI
and
ch p
n
-1/,
1
Ich (nrJ- ) -II}
p ""''''11n
-+00,
stand for the largest and smallest characteristic
on the other hand, by Lemma 2,2, under Ho: f? = Q,
I.
I
11
(2,32)
Also, note that
E{n
-1 U'kr ~l U k } = E{-l
-1 kU1·k] }
n Trace [r U
"'11 . "'11
"'11 "'11
~
= E{n
-1
~
Trace [r
~
=pn-1I~=1
~pk/n,
-1·}
J"'11.
k]
=n
-l\'k
n-i
-1
L' 1 - - .-1 E (Trace [r s. ] )
1= n - 1 +
~
"'111
[as ES
"'111. = r. V 1
(n - i)/(n - i + 1)
V 1
~.
~
i
~
n - 1]
~k ~n.
Therefore, by (2.32), (2.33) and Theorem 2.1 of Birnbaum and Marshall
(1961) for any
(2.34)
"-
P
~
a
n2
~
. .. a
nn
}, 0 < 0
I(
1
'
e: > 0 •
I e: }
max a k In -1 U'k I -1 Uk>
1sks[no] n
"'11
"'11
{
~ pn
Thus, i f
{a
nl
-1
(anl +···+ an[no])/e:
q={q(t), tEE}
continuous function of
(2.35)
r
t
be a non-negative, and non-decreasing and
such that
[q(t)]-2dt <00,
0
-1 2
2
(-2
then n (q (lin) + ... +q ([no]/n)) S J [q(t)] dt, V 0 <0 sl, From
O
(2.35), it follows that by choosing 0(>0) sufficiently small, the
o
righthand side of (2.34) can be made small when
On the other hand, for
o
~
t
~ 1,
a. = q-2 (i/n),
n1
t >0(> 0), q -2 (t) Sq -2 (0) <00,
the weak convergence of
~
i
~ 1.
so that for
in the uniform (or Skorokhod
metric) insures the same under the "sup norm" metric
(2.36)
j
P (x, y)= sup{lx(t) -y(t)l/q(tl,
q
te:E},
Thus, from Theorem 2,1, (2.31), (2,34), and (2.35), we arrive at the
following.
12
The weak convergence in Theorem 2.t holds in the sup
Theorem 2.6.
nom metric
when (2.35) ho Zds •
p
q
Both of these theorems are useful for the proposed RST procedures.
3,
WEAK CONVERGENCE IN THE DISCRETE TIME MODEL
•
As in Section 1, we conceive of
where
R
k
o = to < t
has
r
k
(3.1)
(3.2)
s.
J
l
2
~
~
, k
k
R
m
1,
j =0, I, ••• , m.
failures in the interval
D. (j
J
n~. =Lk>.n. , for
n .. =I(Y. £0.),
1J
1
J
t
::> ••• ::>
and we denote by
D. = [t .• t. 1)'
J
J' J+
Let there be
R ::- R
subjects whose survival times are
< ••• < t . < t
= 00
m m+l
l
m risk sets
1J
-J 1k
~O).
Also, let
O~j ~m,
i=l, ... ,n.
Then, proceeding as in Section 6 of Cox (1972), we obtain the derivatives
of the partial log-likelihood functions (evaluated at
U*k=li'~
1{1i'~
l(n1J
..
"11
l.J=
l.1=
(3.3)
*
:lnk
(3.4)
Ii' k
= l.j =1
-n~.s./r.)z.},
k=l,
1J J J ~1
s.(r. -s.) n
J J J Ii'
r. (r. - 1) . ~ (.tinij
J J
1-1
=L~J= 1 (s.J (r.J
.@,
- s -llr.) S* . ,
J
J "11J
= Q)
as
e"
.. .,m,
...'
-*
*-*
-.tj H.tin ij - kj) ,
say,
where
-=*
z.
(3.5)
~J
In
* )/r.,
= ( . 1z.n..
1= "'1 1J
J
[The close relationship between
need not be overemphasized.]
sl'
• , 'J
s.
J
m
B~O = B(~l'
r , ... , r ,
m
l
•• •, Z ,
"11
sl'
in (2.4) and
In this case, we let
n .. , j
1J
•, •, S
~k
J' =1, ... ,m.
,
m
~k,
1 :s: i :S:n)
r , .•• , r ) .
l
m
for
S*k
"'n
in (3.4)
B*k
.•. , "11
Z ,
n = B(Zl'
~
k = 1,
I
••
,
rn,
while
Then, by arguments
very similar to those in Lemma 2,2, we arrive at the following
Lemma 3, l.
Under (2. 7J and HO:
~ = Q.,
for every n,
{~k' B~k' k :s:m}
is a martingaZe.
We may also note that by arguments very similar to those in
I
13
Section 2, under (2.7) and
~
HO:
= Q,
'\k
n -1 (J * k -/"_lSj
(r.
II
l~k~m
"1l
J~
J
max
(3,6)
- s.)r.~ 1[)
J
II
P
.--+ 01
J
as
n
-+0",
Now, to study the desired weak convergence results, we consider two
•
different cases:
(I)
is fixed, so that as .n
m
for every
J
both increase
j (= 1, •. " m), This situation arises when we have a given
number of ordered categories, while
(II)
and
s.
-+00,
m =m(n)
-+00
but
max
n
s./n
l~j ~m(n) J
arises when the width of D. (j
~
J
is large.
-+0,
as
n
This situation
-+00.
is small, so that there are a
1)
large number of cells and the possibility of ties is no longer
negligible.
In case (I), by virtue of
uncorrelated.
(3.7)
n
B~k_l'
Moreover, given
-~
*
*
(U
k - "'l1
U k - 1) =n
"'l1
is generated by the
(over the set
Lemma 3.1,
rkl
* - ~k-l'
*
~k
k
~
the conditional distribution of
-~.n
-1
l'_lZ'
(n'1 k - skrk
1- ......1
m1 k )
,
equally likely realizations of the
Rk = {i: nik = I})
are all
1
n
ik
and by an appeal to the classical
permutational central limit theorem [c.f. Hajek (1961)], we conclude
that this conditional distribution is asymptotically (in probability)
multinormal with null mean vector and dispersion matrix
(3.8)
. (k
Thus, using a chain of conditioning
some routine steps that given
t
of
* - ~k-l) ,
{n -~ (!!nk
B~o,
1 ~ k ~m}
=1,
" ' , m) ,
(k =m -1, ••• ,1),
i f follows by
the joint conditional distribution
is asymptotically (in probability)
multinormal with null mean vector and dispersion matrix which is
block-diagonal with the matrices
b(k)'
1 ~k ~m
in (3.8),
This, in
14
(2.29) that for every
Iln~1
(3.9)
men)
iL
£
>0
E{~k -~k-l)(~k -~k_I)'I(II!:!nk -!:!nk.l"
>£Iil)IBnk _l }
Lo,
•
I
so that by Lemma 3.1, (3.. 6) and (3.9L the invariance principle for
martingales, developed in Scott (1973), holds for the multivariate case
treated here.
The result is parallel to Theorem 2.1, where we replace
~k' ~k and:Ink
4.
by
~k' ~k and ~k' respectively.
INVARIANCE PRINCIPLES UNDER LOCAL ALTERNATIVES
With a view to study the asymptotic power properties of the RST
(to be considered in Section 5), we proceed now to extend the results
HO: g =Q may not hold, We conceive of a
of local (Pitman-type) alternative hypotheses, where
of Sections 2 and 3, when
sequence
{K }
n
g = B~(n) .= n -~~, for some A £RP
n
'"
'
and desire to study the weak convergence results under
(4.1)
K :
This
n
task is greatly simplified by employing the concept of contiguity of
probability measures
Pn
v
[as in Chapter 6 of Hajek and Sidak (1967)].
be the joint distribution of
~
{K }.
(Y ", Z.),
1
~
i
=Q and P*n be the same under kn . As
Let
=1, "" n under
a first step, we like to
show that under (1.1), (2,7), and (4.1),
{p~} is contiguous to
(4.2)
Let
which
HO(X)
n
be a non-decreasing and non.negative function for
(d/dx)HO(x)
=hO(~) , defined in (1,1), and let g (!) be the
joint density function of
(Y,~)
{p } ,
is given by
~.
Then, by (1,1), the joint density of
*
15
fey, ~; ~) =g(~)ho(y)exp{~'~-e~·~L
"CIO(y)}
(4,3)
13'z
=f(y, ~; Q).exp{~'!+HO(Y)C1-e~~)L
Thus, for testing the simple null hypothesis
~
H :
O
= Q vs. K :
n
t:l.
~
=n
.
~
>..
",'
the (log-) likelihood ratio statistic is
n
(4.4)
10gLn =\.
11og{f(Y.Z.;
n
1.. =
1~1
1
-~.
>")/f(Y.,
Zi;
'"
1
~
0)
~
~ n
~
= n - 2\. 1{>.. ' Z. + H (Y . ) n 2(l - e
/.1= '" "'1
o 1
n- 2>..Z
~i
)}
where we note that
(4.5)
~
n
n 2(1-e
-~>"'Z.
~ ~J..)
1 1
2
=->"'Z. --n-~(A'Z.)
""1
2
"" ""1
t"oJ
1 -~
2
-~
+ -2 n 2 (>.. ' Z.) [1 - exp (en 2>.., Z. ) ]
"" ""1
-..
\
and by (2.28), the last term on the rhs of (4.5) is
uniformly in
(4.7)
i:
10gL
n
~i ~n.
1
(0 < e < 1)
"" ""1
0
-~
P
2
(n 2)(>..,Z.) ,
~
"'1
Thus, by (4.5) and (4,6),
=n-~t1= l(>,,'Z.)[l-HO(Y')]
'" "'1
1
l t 1(>..'Z.)2 + on).
--2
n 1=
~
~
"'1
Now, by the Kintchine law of large numbers, under (2,7),
(4.8)
Al so, under
H :
O
= Q,
~
Y and
~
are independent with
f
(4.9)
E[Ho(y)]r = [Ho(y)]rexp(-Ho(y))dHo(y) =r!,
Thus,
E(~'~i)[l -HO(Y i )]
=
(~'lA)
0 =0
and
V r = 1,2, •. ,
E[(~'~i)(l -HO(y i ))2] =~'I.6.
Hence, from (4.7), (4.8), (4.9), and the above, we conclude that under
H :
O
~ = Q,
variance
10gL
n
.6'I.6.
is asymptotically normal with mean
-
~.6' 11
and
By virtue of LeCam's third lemma [c,f. H~jek and Sidak
(1967, p. 208)], this ensures that
{P*}
n
is contiguous to
{p}.
n
By using (4,3), we obtain by some simple steps that the conditional
density of
~
given
Y = y,
under
K,
n
is given by
16
!-<
!-<
g(~ly) =f(y~ f; ~=n-2~)lff(y~ ~; ~=n"2~)df
(4.10)
= g C.~)
so that under
K
I1
+ n-
!-<
2 (1
- H (y)) (~ O
1\)
'~ + 0 (n '"'
~
E(Z~1. I Y.1 = y) = N11 + n- 2 (I - H (y) ) ~~
fA + 0 (nO
Also, i f
FO(x)
I
~
2) •
be the marginal density of Y (same in both the null
and non-null cases), then
I-FO(y) =exp(-HO(y)),
has the simple exponential distribution,
(4.12)
2) ]
~
n
(4.11)
!-<
[1 -HO(y)] -
so that
Ho(y).
Hence, for any y,
U-FO(y)}-lflO[1-HO(~)dFO(~)
y
oo
={l-FO(y)}-l
fY
(00 -H
[1 -FO(~)dHO(~) =U-FO(y)}-lj, e
(y)
0
Y
dHO(~) =1.
From (2.2), (4.11), and (4.12), we obtain by some routine steps that
under
in (4.1),
{K}
n
(4.13)
n
Thus, if we define
(4.14 )
lim
n-+oo
-1
~n(t)
E{~
(t)
n
k
~a(O
< a < 1)
~
as in (2.11), then
1/
IKn } ~tr'2A
~
~
, V t cE.
Now, under the null hypothesis
independent of Y,
HO: ~ = Q and (1.1), Z is
so that by using Lemma 1 of Bhattacharya (1974),
it follows [as the conditional df of
~
given
y does not depend on
HO' ~l' ""~n are i.i.d.r.v. with the density q(f)·
Hence, by an appeal to Theorem 2,1, the contiguity of {p*} to {p}
y] that under
n
n
and a theorem of Neuhaus and Behnen (1975), we conclude that for any
k:
kin
~a:
with mean
o <a :51, n -~~k is asymptotically normal (under
aI~
and dispersion matrix
employed to show that for finitely many
aI.
(m)
{K })
n
The same technique can be
Points,
!-<
n- 2(U
U)
'""11k l' ••• , '""I1k
m
17
(where
k .In
J
-+ a.
J
has asymptotically (under
:
variate normal distribution with mean vector
dispersion matrix
{K })
a pm
n
and
(a ••••••. am) 0 ...,....,
fA
l
This establishes the convergence of
«(arAa.))
0r,
J
~
in (4.1).
n
As in the proof of Theorem 2 of Sen (1976) , we conclude that the
finite~dimensional
{p*}
n
contiguity of
distributions of
{p }
to
{F,;n}
under
and the tightness of
n
(insured by Theorem 2.1) imply the tightness of
well.
{K }
{F,;n}
{F,;n} under
under
H
O
{K }
n
as
This leads us to the following
Theorem 4.l.
~ +
defined by (2.11) converges weakly to
.
Theorem2.land
~
{Kn } in (4.1) [relating to (1.1)],
Under (2.7) and
£,
where
f
is defined in
!.:
F,;={t:(t)=tr 2 A,
~
'"
,...",
tEE}.
I'J
We conclude this section with the remark that similar weak
convergence results hold for the discrete time case treated in Section 3.
5.
RST PROCEDURES RELATED TO THE COX REGRESSION MODEL
We start with the remark that if one works with U
""11n
in (2.2)
and constructs
r U
nn =U'
""11n""11n""11n'
L
(5.1)
is defined by (2.3), then, under HO: ~ = Q and (2.7) - (2.8) ,
Lnn has asymptotically chi-square distribution with p degrees of
freedom (DF) (by Theorem 2.1) and by Theorem 4.1, under {Kn } in (4.1),
where :Inn
has asymptotically a non-central chi-square distribution with p DF
nn
and the non-centrality parameter
L
~L =~'£~
(5.2)
Now 1 suppose we work with
e.
a sub-vector of
g,
L~,
.
defined by 0.7).
Since
~t
the martingale property in Lemmas 2,2 and 2.3
remain true for these subsequences
{~k} and {~k}
and also
is
18
{J*k
",n L
Lemma 2,4 extends to
{L* }
Thus, the weak convergence of
run
follows on the same line as in the proof of Theorem 2.1, provided
m(= m(n))
-+00
as
n
-+00,
Further, the contiguity being assured by
(4.2), Theorem 4,2 also holds for this subsequence,
claim that under
(2.7)~(2,8)
and
H :
O
~=QJ
chi-square distribution with P DF , provided
As a result, we
L*
has asymptotically
men)
-+00
nm
as
n
-+00.
Also,
if, in addition,
1 .
(5.3)
lim n- men) = '[:
n-+oo
then under
0::; '[ ::; 1 ,
L*tun has asymptotically a non-central
{Kn } in (4.1);
chi-square distribution with p DF
and non-centrality parameter
!.l L* = '[2 (~' r~) = '[2 .!.l L '
(5.4)
Thus, the asymptotic relative efficiency (A.R,E,) of
to
L*nm with respect
Lnn is equal to
(5.5)
In particular, if
'[
is close to
this sense, i f m(n)/n
regression based on
-+1
as
n
1,
-+00,
then (5,5) is also so and in
then the Cox (1972) test for no
L*nm in (1,7) is asymptotically fully efficient -
this result has been obtained for the discrete time model by Efron (1977)
from a somewhat different consideration.
for
'[<1,
However, we may note that
the A.R.E, in (5.5) is less than 1 indicating that too
much of censoring takes away some information,
A common feature of the two tests based on
they are based on the partial likelihood
Lnn and L*nm is that
Lnn and L*run'
respectively,
and, thereby, demand that the experimentation be continued until all
or m of the failures occur,
In the proposed RST procedure, we like
to update the picture as successive failures occur and consider the
following time-sequential procedures,
These procedures are based on
..
19
the partial sequence
{L
nk
~n}
k
,
of partial likelihood ratio statistics,
Note that by (2,10)-(2,11)1
.
~
-1
~
,Lnk (t) = [~(t)]' [:!nn:!nk (t):!nn] [~(t)], V t cE ,
n
n .
(5.6)
where by Lemma 2,4, under (2,7)1 (2,8), and
~
(5.7)
P
-1
~ t
I,
"'P
~
J""1ln"'1l
J k (t)J"'1ln
n
Thus,
L
nk
hand, for
(t)
n 2
behaves like
q (t) =t
-1
,
.§=Q,
H :
O
V teE.
t -1 [~(t)] , [~(t)]
as
(2.35) does not hold and
not behave very smoothly when
t+O;
t-
n
1
-HXl.
On the other
[~(t)]'[~(t)]
does
in fact, by the law of iterated
logarithm for Brownian notions,
lim sup t
t-+O
(5.8)
-1
.
[~(t)]'
[.s(t)] =00,
with probability 1.
Hence, to make statistically informative use of the partial sequence
either we use some appropriate weight function or we
\
start monitoring of the experimentation only when a given number of
failures have already taken place,
This leads us to the following
two types of RST procedures.
(i) Type A RST procedures.
o
k = [fiE:].
n
N =
{
c n,
(5.9)
"
c*
n
and let
c:
Define then the stopping number
min{k (k
where
We choose some
if
0'
n
~ k ~ n)
:
L k > c*},
n
n
Lk~c*,
VkO~k:<::;n.
n
n
n
is a suitable constant, to be defined later on.
operationally, starting at thekO-th
n
failure
o
Y 0'
nk
Then,
we monitor experi-
n
mentation and at each failure Y (k ;::: k ), we compute L : if L
is >
nle
nk
n
nk
c~,
we stop experimentation at that time along with the rejection of
O:
H
~
=Q and if Lnk
~
c*,
n
we continue experimentation until the next
failure occurs and reinspect the process at that failure,
If, we do
20
not stop on or before the n-,th failure, experimentation terminates
along with the acceptance of
allowing
k to range over
and, in that case,
N
E:
HO' In (S,g), we may modify N( by
k0 ~ k ~ m for some predetermined m(~ n)
n
with probability
~m
In theory, it does
I,
not make any difference, but, in most practical problems, we may be
inclined to choose an
m well below
n,
depending on the time and
cost of experimentation (at our disposal).
We need to choose
c*
n
in
such a way that
PUnk >c~
(5.10)
where
a(O < a < 1)
for some
k~ ~k ~nIHo: f? =Q} = a,
k:
is the desired level of significance of the RST.
(ii) TYpe B RST procedupe.
Here, we do not use a constant
{c~k' 1 ~k ~n}
boundary and choose a sequence
of positive numbers
and denote the stopping number by
(5.11)
N* =
min k(l ~k ~n):
{n,
L
>c~k'
nk
L
~c~k' V k ~n .
nk
if
Then, we proceed as in the Type A procedure, but using the variable
boundary instead of the constant
c* .
n
Here, we need to have
(5.12)
[We do not need to wait until the occurrence of y
c*n
PUnk >A,
-I-
n
E
for some
>0
and
say,
under HO:!2, = Q,
A > 0,
k: k~ ~k :s;nl~o}
p{t-l[~(t)]'[~(t)] >A for some
= P (A; E),
n
in (5,9)-(5,10), we now appeal to Theorem 2.1
and conclude that for every
(5.13)
and the process
kO_th failure.]
may terminate even prior to the
To determine
nk
0
E
~t:s;l}
21
TABLE 1
Table for the simulated values of
-
p =1
a=,Ol
.
--
A
a
P =3
P =2
,OS ,10
,01
,OS
(a,
for typical
£
,01
.10
£)
-
,OS
=
P =4
,10
.01
.05
.10
.005
12.11 9,24 7.78
15.42 12,46 10.68
18,19 15.14 13.10
20.97 16.83 14.99
.010
11.97 9,06 7.62
15.24 12.23 10.52
18,10 14.52 12.85
20,44 16.76 14.86
.050
11.49 8,35 6.86
14,85 11,90
9.88
16.99 14,19 12,37
19.72 16,18 13.94
.100
11. 09 7,79 6.35
14,69 11.48
9,39
16,92 13.98 11,87
19,58 15.72 13.46
.150
10.63 7.51 6.10
13.94 10.99
8.97
16.75 13.60 11.47
18.62 15.24 13.18
.200
10.48 7.34 5.90
13.92 10.55
8.87
16.50 13.02 11.21
18.10 14.96 12.96
Aa£ be the solution of peA;
Thus, if
{c*}
n
£) =a,
any sequence of positive numbers for which
then we may choose
c* +A
n
a£
Aa£ are not yet available.
Analytical solutions for
as
n +00.
However, certain
simulation studies are made by Majumdar and Sen (1977) and Majundar (1978).
We report above, in Table 1, some of these simulated values for
c~k
For (5.11) - (5 .12), we take
= (n/k) c~,
1
~ k ~ n,
(5.6), (5.7), (5.12), and Theorem 2.1, we have for
P{Lnk > c~k'
(5.14)
+
Thus, if
A~
{*Lnk , 1 ~k ~n}]
p >1
k ~n IH O}
p{ [~(t)] , [~(t)] > A,
for some
be the solution for
[For this choice of
for
for some
(for
p
c~k'
Pt =a,
4.
'
=Pt '
say.
then we have choose
c* = A* .
n
a
we really work with the partial sequence
Again, analytical solutions for
=1,
tEE}
~
so that by
c* +A
n
p
A~
are not known
this is, however, known), and, we quote the
following simulated values obtained by Majumdar and Sen (1978).
By virtue of Theorem 4,1, we are also able to express the
asymptotic power of these RST procedures interms of the boundary crossing
probabilities of
{~(t),
t EEl.
From (5.9), (5.10), and Theorem 4.1,
22
TABLE 2
Table for the simulated
values of A* for some typical
(a, p)
a
A*
a
a
P
la
,01
,05
,10
7.90
5,02
3.84
2
10.37
7.29
5.52
3
13.76
9.30
7.73
4
15.13-
10,96
9.24
a)Exact values.
we have that under
{K}
in (4.1), the asymptotic power of the Type A
n
RST procedure is equal to
(5.15)
Similarly, for the Type B RST procedure, the asymptotic power under
{Kn } in (4.1) (when we choose
c~k
!-.
= (n/k) c~, k $ n) is given by
!-.
p{(f,;(t) +tr 2 A)'(f,;(t) +tr 2 A) >A*ex.
(5.16)
,.....,
I"'OV
f'Jf'OJ
for some
tEE},
,....,,....,
A comparison of (5.15) and (5.16) demands extensive numerical integra-
tion and simulation studies.
So far, we have considered RST procedures based on
It
{L~k'
is also possible to use
k ~m}
{L
nk
when m-+00 with
, k ~n}.
n -+00.
The stopping variables are similar to those in (5.9) and (5.11) and
when
min -+ 1,
min -+T,
(5.13) and (5.14) also stand valid.
for some
the range of
T < 1,
t(E~t~l
However, if
then in (5.13) or (5,14), we have to replace
or
tEE)
by
E~t~T
similar changes are needed in (5.15)-(5.16),
(or
tEE[O, TD
For the discrete time
parameter case, treated in Section 3, when m =m(n) -+00 with
(i.e"
case II), the same results apply.
and
In case I (i.e.,
n-+ oo
m fixed),
23
~n(t)~
instead of the process
~n(tj)'
t j EE,
j =1, ""m,
t EEl
we have only finitely many
As such, t;he useof\).E or
A~
(5.13) or (5,14) will result in a more conservative test i,e"
actual level of significance may be much below the specified
in
the
a.
However, as in Majumdar and Sen (1977), this will continue to be a nice
test.
Finally, throughout the paper, in the usual fashion of an
analysis of covariance model, we have taken
the other hand, the
Z.
~1
Z.
~1
as stochastic.
If, on
are non-stochastic, the problem becomes a
little more simple and a direct multivariate extension of Sen (1975)
yields the desired results.
REFERENCES
BHATTACHARYA, P.K. (1974). Convergence of sample paths of normalized sums
of induced order statistics. Ann. Statist. 2,1034-1039.
J
BIRNBAUM, Z.W. and MARSHALL, A.W. (1961). Some multivariate Chebyshev inequalities with extensions to continuous parameter processes. Ann. Nlth.
Statist. 32, 687-703.
CHATTERJEE, S.K. and SEN, P.K. (1973). Nonparametric testing under progressive
censoring. Calcutta Statist. ASBO. Bull. 22, 13-50.
COX, D.R. (1972). Regression models and life tables. Jour. l?oy. Statist. Soc.
Serf B, 34, 187-220.
COX, D.R. (1975). Partial likelihood. Biometrika 62, 269-276.
EFRON, B. (1977). The efficiency of Cox's likelihood function for censored
data. Jour. Amer. Statist. Asso. 72,557-565.
"
HAJEK, J. (1961). Some extensions of the Wa1d-Wo1fowitz-Noether theorem.
Ann. Math. Statist. 32,506-523.
/
"
"
HAJEK, J. and SIDAK, Z. (1967).Theory of Rank Tests. Academic Press, New York.
HOEFFDING, W. (1948). A class of statistics with asymptotically normal
distribution. Ann. Math. Statist. 19,293-325.
MAJUMDAR, H. (1978). Rank order tests for multiple regression for grouped data
under progressive censoring. Calcutta Statist. Asso. Bull. 26,1-16.
MAJUMDAR, H. and SEN, P.K. (1977). Rank order tests for grouped data under
progressive censoring. Comm. Statist. Ser A. 6, 507-524.
24
MAJUMDAR, H. and SEN, P.K. (1978). Nonparametric tests for multiple regression
under progressive censoring. Jour. MUZtivar. AnaZ. 8, 73-95.
NEUHAUS, G. and BEHNEN, K. (1975). A central limit theorem under contiguous
alternatives. Ann. Statist. J, 1349-1353.
SCOTT, D.J. (1973). Central limit theorems for martingales and for processes
with stationary increments using a Skorokhod representation approach.
Adv. Appl. Probability 5, 119-137.
SEN, P.K. (1976). Weak convergence of progressively censored likelihood ratio
statistics and its role in asymptotic theory of life testing. Ann.
Statist. 4, 1247-1257'.
SEN, P.K. (1979). Weak convergence of some quantile processes arising in
progressively censored tests. Ann. Statist. 7,
t