910 Part B

•
A NOTE ON STEP-DOWN PROCEDURE IN MANOVA
PROBLEM WITH UNEQUAL DISPERSION MATRICES
By
S. R. Chakravarti
Department of Biostatistics
University of North Carolina at Chapel Hill
and
University of Calcutta, India
Institute of Statistics Mimeo Series No. 910
,
"
FEBRUARY 1974
•
A NOTE ON STEP-DOWN PROCEDURE IN MANOVA
PROBLEM WITH UNEQUAL
D~SPERSION ~TRICESI
By S. R. Chakravorti 2
University of North Carolina, Chapel Hill
In this paper we have considered step-down procedure
in multivariate analysis of variance problem when dispersion
matrices are different and unknown.
The distribution
problem of the test criterion has also been studied under
the null hypothesis.
1.
Introduction.
'\IVV\fV\/\.IV\
Step-down procedure in standard MANOVA problem has
been considered by J. Roy [7].
The essential feature of this procedure
is that if on some a priori grounds the variates are arranged in
descending order of importance, then the test procedure can be carried
out sequentially by considering marginal and conditional distributions
of the variates concerned.
At each stage F-statistic can be used which
are independently distributed under the null hypothesis so that the
overall hypothesis can be tested by combining the component tests.
Optimum properties of this procedure have been discussed by Roy [7]
and Roy et a1. [9].
lWork sponsored by the Aerospace Research Laboratory, U. S. Air Force
Systems Command, Contract F336l5-71-C-1927. Reproduction in whole or
in part permitted for any purpose of the U. S. Government.
20n leave of absence from the University of Calcutta, India.
•
2
In this article the problem has been considered when the dispersion matrices are not equal.
It has been shown that by transforming
•
the original vector variables to Scheffe-vector-variab1es (Eqn. (2.2)
below) Hote11ing's T2 test can be constructed at each stage of the stepdown procedure and these T2 ,s are shown to be independently distributed.
It may be noted that the distributions of the Scheffe-vector-variab1es
chosen from the original vector variables at different stages (vide,
Remark at the end of Sec. 2) may not, in general, be the same as those
considered earlier.
This is possibly true under certain restrictions
(to be shown in Sec. 3), when the test constructions will be difficult
in this situation.
However, a reasonably satisfactory solution can be
attained only in Bhargava's [3] procedure, which is a
n~dification
of
Anderson's [2] procedure in generalized multivariate Behrens-Fisher
problem.
2.1.
Hypothesis and Test Construction.
Let x(t) (lxp) be the a-th
-a
observation vector in t-th population (a=l, ••. ,n , t=l, .•. ,m) and
t
distributed as N (n(t), E ), where n(t) (lxp) is the mean vector and
p -t
~t(pxp) the dispersion matrix in t-th population.
The problem is to
test
(2.1)
against the alternative of at least one inequality among n(t),s.
Anderson [2] proposed a Rotel1ing's T2 test for (2.1) by assuming
n l < •••
~ nm and transforming the variables ~~t) to SCheffe-vector-variabletlll
3
•
(2.2)
for r=2, ••. ,m, a=l, ••• ,n , where U = (u(2) , .•. ,u(m»
-a
-a
-a
l
jointly follow a
p(m-l)-variate normal distribution.
Now suppose the p variables in the vector x(t) are arranged in
-a
descending order of importance and the ordering remains the same for
each t=l, ... ,m and is given by
(t)
(2.3)
= (xl
a
, ••• ,x
(t)
pa
)
After transforming the vectors x(t) to (2.2) if we order the variables
-a
in u(r) (lXp) as (Ul(r) , ••. ,u(r», these new variables correspond to the
-a
a
pa
ordered variables in (2.3) through this transformation.
!!a(lXp(m-l»
= (U , .•• ,U ), where U.
-la
-pa
-la
(2)
Now writing
(m)
(U . , ..• , U. ), we have the
la
la
distribution of U as N ( 1)(8,f), where for r=2, ..• ,m, i=l, ••• ,p
-a
p m-(2.4)
[(p(m-l)Xp(m-l»
(2.5)
= (Ii.)' I· . (m-lxm-l) =
J.~
n
(2)
n
a~:)J
~ (m)
l
+ diag ( -l a.. , ... , a . )
n 2 1J
nm i J
where i,j=l, •.. ,p,
~(m-lxm-l)
a~:) is the (i,j)th element of Et , t=l, ••. ,m, and
1J
is the matrix with all the elements unity.
Now if we consider j-th step-down 9rocedure, then we are to
consider the conditional distribution of U. (lxm-l) for fixed
-Ja
U(. 1)
- J-
a
= (U l , .•• ,U. 1 ), which is an (m-l)-variate normal distribu- a
-J- a
tion with mean vector W. and residual dispersion matrix f. 1
-J
where
. l'
-J. "",J-
•
4
•
(2.6)
~J~-l
(2.7)
=
(13~
. 1)
-J, 1,···,13~
-J,J-
is a matrix of order (m-1)x(m-1) and B. 1 is the (j-1)th
where I3
- j ,s
-J-
order step-down regression matrix.
Under this set up, the hypothesis (2.1) can be written as
follows
.
(2.8) H [n
(1)
= .•• =n
o -
-
(m)
]
~
n
H :[8=0]
0
I
H(j ) [8 . =0 8.
-.
--
J=l
=0 ]
-J - -(J-1) -
0
-
Thus the component hypothesis R(j)[s.=O] can be tested from model (2.6)
o
-J2
by Rote11ing's T (Anderson [1], Page 187), where
A -lA
(n -(m-1) (j-1)-1)n s .S s~,
1 -J - U.-J
1
(2.9)
J
where
A,
A,
(13.
. . 1)
-J, 1,···,13
-J,Jand
n
~U
This
j
=
~U.
J
1
a~l (~ja-§j-~(j-1)a~j-1)'(~ja-§j-~(j-1)a~j-1)
is distributed as Wm_1 (n 1-(m-1)j, r.
1
. 1) and
. . . J. , ... ,J-
(n -(m-1)j) (m-1)
1
-1
. -1
(n -(m-1) (j-1)-1)
1
F(m-1,n -(m-1)j), j=1,2, ... ,p.
1
2
T is distributed as
j
•
5
~u.
T2 ,
1
It is clear that for fixed j,
are independently distributed (Anderson [1]).
J
•••
•
A
2
Independence of Tl,
..• ,T p2 .
2.2.
~.
~J
and
To prove that
,T 2 are independently distributed under H , let us consider a vector
p
.
0
~(lxm-l) of real elements so that ~~u.~'/~[j.l,•.• ,j-l~' is distributed
J
as xj(nl-(m-l)j).
this
x~
not.
J
Now following Roy et ale [9], Page 47, it follows that
is distributed independently of U(. l)t', whether H(j) is true or
- J-
-
0
So that X~, ... ,x~ are independently distributed.
therefore, (Rao [61, Page 453) that
distributed.
Also under H(j)
o
'
t.t'
~J~
~U
It follows,
for j=l, •.• ,p are independently
j
is distributed independently of
A
u(.
1) -t' and is true for every
~ Jof "Q(j-l)'
(2.10)
~J
so that
~.
~J
is distributed independently
Since T~ in (2.9) can be written
J
[nl-(m-l) (j-l)-l]
where the distribution of
e.
~,
(Rao [6], Page 458).
-1 2
A
A
-1
A
-IA
A
-1
A
J
-IA
A
-1
~.SU ~~/~.r.
~J-
A
nl~·r.
1 , ••• ,J. l~~(~'Su
~~/~.r. 1 , ..• ,J. l~~)
-J~J'
~J ~J
.~J ~J~J'
~J
Tj
.~J
-J~J.
1
A
. l~~ does not depend on
, ••• ,J- ~J
HenceJunder H(j), the conditional distribution
0
of T: for fixed U(. 1) does not depend on U(. 1) (j=2, ••• , p) and T12
J
~ Ja
- J- a
is distributed as (nl-m+l)-l(m-l)F-distribution with d.f. (m-l, nl-m+l).
Hence unconditionally also T~, ... ,T~ are independently distributed.
The test criterion for the overall hypothesis (2.8) can be constructed from the component tests either by using union-intersection
principle (Roy [8]) or by considering the test criterion A =
~ A(j) ,
j=l
is the product of
(m-l) independent beta variables.
The exact as well as asymptotic null-
distribution of the statistic A are available (Chakravorti [4], eqn.
4.39 and 4.55 for zero non-centrality parameter).
•
6
REMARK.
It may be noted that if we start with the marginal and con-
'
1 d'1str1'b ut i ons 0 f xl(t) ' x ( t )g1ven
.
(t),
di t10na
xl(t)
a
a , .•• ,x pa g1ven
2a
xl(t) , ••• ,x(t ) from the distribution of x(t), and then choose the
a
p- l a
.
-a
Scheffe-variables in each stage of the step-down procedure, the distributions of the resulting vector-variables will not be the same as
those obtained from (2.2).
Since these variables are correlated both
variate-wise and group-wise, we are to impose a number of restrictions
on the regression coefficient matrix B. 1 to satisfy this requirement,
-Jin which case a satisfactory solution of the test construction is not
easily available by Anderson's procedure.
However, if we assume that the regression coefficients of
x~t)
J
on x(t) (s=1,2, ••• ,j-l) are same for t=1,2, ••. ,m, we can apply the
s
procedure of Bhargava [3], which is the modification of Anderson's
procedure.
3.
Following Bhargava [3] let us assume that
nl=(m-l)n', n = .•• =n =n' and consider the following transformation on
2
m
x(t)
-a
'
= x(r)_x(l)
-a
-a+(r-z)n' ,
(3.1)
r=2, ... ,m; a=l, •.• ,n'.
Then under the set up considered in Sec. 2.1, we can choose the vector
l!a(l Xp(m-1»
= (U l , .•. ,U ), which is distrubuted as N ( 1)(8,r*),
- a
-pa
p m--
where
(3.2)
(2)
(m)
8(r)_ (r)_ (1)
8= (8 , ... ,8 ); 8, = (8 , , ... ,8 . ) ; . - n.
n,
- -1
-p
-1
1
1
1
1
.1
(3.3)
r*=(t~,);
t~,=a~:)I
1 + diag(a~:) , ••. ,a~~»; i,j=l, •.. ,p.
1J -m1J
1J
t
t
1J
.
1J
Then j-th step-down procedure will lead to the model (2.6), where 8. 's
-J
•
7
are defined according to (3.2) and submatrices of B. 1 given by (2.7)
- J-
are for s=1, ••• ,j-1
(3.4)
e.
-J,s
•
(m-1 m-1) = diag(S~2,2) , ••• ,S~m,m».
.
J,s
J,s
When S~r,r) remains constant for r=2, ••• ,m,(Sj (say» the conJ,s
,s
ditions stated in the remark of Sec. 2 are satisfied. Now since the
components of U. = (U (2) , ••• ,u~m»
ja
-Ja
Ja
are independently distributed let us
write the null hypothesis (2.8) as follows from (3.2),
(3.5)
m (r) (r)
(r)
(r)
\--}
~
(l HoJ' [8 J. =ol~(J'-l)=Q]; ..§(J·-1)=(8 1 ,···,8 J·_1 )
j=l r=2
p
- n
Thus we can construct test for H(:) by using individual estimates to
oJ
m
S.
for r=2, ••• ,m and hence the combined test for H • =
H (:) by
J,s
OJ
r=2 OJ
n
(3.6)
L
j
=
m
II
r=2
[l+j(ll-j)-l E ]-1
jr
where Fjr fo110wsCeMMJ . f-distribution with (j;"~Jd.f. Obviously
L. is the product of (m-1) independent beta variables, the distribution
J
of which is well-known (Anderson [1], Rao [5]).
Test criterion of the
p
overall hypothesis can therefore be constructed by considering L = II L .•
j=l J
The independence of L. for j=l, ••• ,p can be verified by the similar
J
arguments as in Sec. 2.2.
~.
The author is grateful to Professor P. K. Sen for
helpful discussion.
•
8
•
REFERENCES
[1]
ANDERSON, T.W. (1958).
John Wiley & Sons, New York.
Analysis.
[2]
An Introduction to Multivariate Statistical
ANDERSON, T.W. (1963).
A test for equality of means when
covariance matrices are unequal.
Ann. Math. Statist. 'VV
34
671-672.
[3]
BHARGAVA, R.P. (1971).
A test for equality of means of multi-
variate normal distributions when covariance matrices are
Cal. Stat. Assoc. Bulletin 20 153-156.
unequal.
[4]
'VV
CHAKRAVORTI, S.R. (1973).
On some tests of growth curve model
under Behrens-Fisher situation.
Institute of Statistics
Mimeo Series No. 870, University of North Carolina,
Chapel Hill.
[5]
RAO, C.R. (1951).
An asymptotic expansion of the distribution
of Wilks' A-criterion.
Bull. Inst. Internat. Stat.
~
Part II 177-180.
[6]
RAO, C.R. (1965).
tions.
[7]
Linear Statistical Inference and Its Applica-
John Wiley & Sons, New York.
ROY, J. (1958).
Step-down procedure in multivariate analysis.
Ann. Math. Statist.
[8]
ROY, S.N. (1953).
~
1177-1187.
On a heuristic method of test construction
and its use in multivariate analysis.
kit
[9]
Ann. Math. Statist.
220-238.
ROY, S.N., GNANADESlKAN, R., SRIVASTAVA, J.N. (1971).
Analysis
and Design of Certain Quantitative Multiresponse Experiments.
Pergamon Press, New York.
•