QUADRATIC UNBIASED ESTIMATION OF VARIANCE COMEONENTS IN LINEAR IDDELS HITH AN EMPHASIS ON TEE ONE-HAY CIASSIFICATION by James Howard Goodnight Institute of Statistics Mimeo Series No. 850 November 1972 iv TABLE OF CONTENTS Page 1. INTRODUCTION....... 1 2. STATEMENT OF 'rtfE PROBLEM • • 4 2.1 2.2 2.3 2.4 2·5 2.6 3. COMPUTING LOCALLY BEST QUADRATIC UNBIASED ESTIMATORS • 3.1 3.2 3·3 3.4 4. The Mathematical Model Unbiasedness . • . • . Invariance Concepts • • • • • • • Locally Best Quadratic Unbiased Estimators Locally Best Quadratic Unbiased Estimators With Addi tional In-v-ariance Restrictions . . . . . A Suggestion For Obtaining Invariant Quadratic Unbiased Estimators • • • • • • • • • • • • • Forming The Matrices Needed • Computing The A Matrix Computing The BA Matrix Computing Other Invariant Estimators QUADRATIC UNBIASED ESTIMATORS FOR THE ONE-WAY CLASSIFICATION 4.1 P '1.'''''Tor'd • • • • • • • • • • • • • • • 4.2 The Mathematical Model • • • • • • • • 4.3 Invariant Quadratic Unbiased Estimators For The Variance Components • • • . . . . . . 4.4 Re.lative Efficiencies • • • • • • • . • 4.5 Choosing Reliable Estimates With Limited Prior Knowledge • • • • • • • • • • • • • • • • • • 4.6 Comparison of Invariant Quadratic Unbiased Estimators To The Standard Analysis of Variance Estimators • • • • • • • . • . . • • . • • • 5. 4 4 5 8 10 15 15 19 21 21 23 23 24 25 29 36 40 SUMMARY AND SUGGESTIONS FOR FUTURE RESEARCH 5.1 Summ.ary.. • • • • • • • • • 5.2 Suggestions For Future Research • • • • . 6. LIST OF REFERENCES • 7. APPENDIX • • • • • ......... 7.1 An Algorithm For Factoring A Symmetric Positive Definite Matrix in Place • • . • . . • • 7.2 An Algorithm For The Inversion Of An Upper Triangular Matrix in Place • . • . • •• •• • 7.3 An In Place Generalized Inverse Sweep Algorithm • 1+5 46 50 51 52 53 nrrRODUCTION 1. The classical approach to the unbiased estimation of -'rariance components for -~balanced data is one of choosing several different quadratic functions of the data, equating them to their expected values, and sol~~ng the resultant system of equations. There are, of course, infinitely many quadratic functions available for equating observed to expected values, and solving to provide unbiased estimation. However, much of the previous work in this area centers around quadratic forms which bear analogy to those used with balanced data; in particular, much work has been done using the methods outlined by Henderson [1953J. Rao [1972J states that the classical methods lack a clear theoretical basis and that the classical procedures are: intui tion ll • "ad hoc and much seems to depend on Numerous authors 1 have, in fact, compared two or more of the classical estimators to determine which, if any, has the smaller variance for a particular design. However, the dedsion as to which estimator among a given set is "best", always seems to depend upon the unknown values of the components being estim.ated. Hence, the idea of achieving a uniformly "best" unbiased variance component estimator in the "generally" unbalanced situation 2 appears to be improbable. III 1 Review articles by Searle [1971J ~nd Harville [1969bJ describe much of the earlier work in variance component estimatioIl and give some of the important references. 2 Situations s~ch as those that arise from the loss of data, or lack of data, etc. are implied here. Conceivably, specific designs wi th planned imbalanced may be found which yield uniformly "best" estimators. Rao [1971bJ provides some necessary cond.i tiOD.B for this to occur. -e 2 fact, Read [1961J has proven that there exists no ql~adratic E;stimator of' the "between!' component in the unbalanced one-way classification for which the variance (assuming normality) is uniformly smaller than that of every other quadratic estimator. Recognizing, perhaps, the fundamental difficulties which arise from the classical approaches to variance component estimation, authors of recent papers have focused their attention on choosing quadratic forms with some sort~; of o..p timal properties. Ha~viJ.le [1')69aJ, f,:;,1' example, has used thc:~ results of HUltquist aLld Graybill [1965J on minimal suffident statistics in conjunction w'i th Koch IS [1967J J.emma on the variance of quadratic forms to establish the basic form that the matrix of a quadratic form should have when it is to be used for estimating the components in a one-way classification. Townsend [1968J gives locally best quadratic unbiased estimators for the variance components associated with the one-way classification with zero mean. Recently published papers by Rao [1970, 1971a, 1971b, 1972J outline new techniques referred to as MINQUE (Minimum Norm Quadratic Unbia.sed Estimation) and MIVQUE (Minimum Variance Quadratic Unbiased Estimation). Although Rao develops MINQUE and MIVQUE with fairJ~ relaxed assumptions on the distributional properties of the random effects involved, he does consider the special case where the random effects are normally distributed. In this case ML~QUE aQd MIVQllli estimators are the same and provide locally Ilbest" quadratic unbiased estimators. r,rhe present paper is restricted to the quadratic unbiased estimation of the variance components in linear models for which the random effects are taken to be norma.lly distributed. The basic: -e 3 objectives of this paper are to extend MIVQUE theory tc provide estimators whose variance is functionally dependent on as few of the unknown parameters as possible, t.o provide computational techniques for MIVQUE and its extensions, and to apply MIVQUE techniques to the unbalanced one-way classification in order to develop an estimator for which no ~ priori knowledge about the variance components is necessary, yet the efficiency of which is in some sense optimal. 4 2. STATEMENT OF THE PROBLEM 2.1 Let the NX 1 The Mathematical Model vector of random variables Y have the linear structure (2.1) (i ::: 0, •.. , m) where X. (with Xm:::~)' J. is an 13 0 is an nO parameters, and each 13. N X n. J. X 1 vector of unknown non-stochastic (i::: 1, ..• , m) J. matrix of given values is an n. X 1 J. vector of uncorrelated random variables assumed to be normally distributed with 2 a-.1 mean zero and variance (i ! j) l n. Furthermore, each J. are assumed to be uncorrelated. 13 i and From the above it follows that Y fOV Normal (Xa!3o' V) , where 2.2 Unbiasedness The problem of quadratic unbiased estimation of the variance components ... , such that the • any quadratic form in a- 2 m is one of choosing quadratic forms 2 ::: a-. l (i ::: 1, ... , m) . I Y Q.Y l The expectation of Y, as defined by (2.1), is 5 2 Thus a quadratic unbiased estimatoi' of a matrix such that Q.1 1 1, ... , m) and if i = j } if i F j 2.3 (i cr. for 1, ... , m • j Invariance Concepts % By imposing additional restrictions on I other desirable properties of the estimator The concept of I!invariance (i ~Y Y may be obtained. On the translation of the t3 t3 0 for any choice of values for the vector t3 * 0 condi tion for invariance with respect to t3 X Q type of = parametel,11 I is said to be lIinvariant on the translation of the t3*0 0 A quadratic form Y QY is considered by Rao, Harville, and others. condition is also necessaryl unless ... , m) = 1, is 0 parameter ll if A sufficient I o =0 This The appeal for this 0 . t30-invariance can be seen in that recoding of the data by subtracting * Xd30 does not alter the estimate for a particular effect. In addition, quadratic forms in contain any elements of Po B ' ~lf O Y, (2.1), as defined by do not in their variance if they are t3 o- invariant. The variance of a symmetric quadratic form l Var (y QY) • • Y/QY 2 . ~ tr QVQV + 4f3if.;QVQX.Jo lJ is (2.4) 1 Expansion of the right hand side of (2.3) with the definition of Y from (2.1) involves the term X~QX.m or for (2.3) to hold for any non-trivial X/QI 0 * t3 0 . which must be zero 6 Applying the definition of V given in (2.2), the above variance may be expressed as Furthermore, by defining, for any real matrix ssq(A) A, the matrix operator to be equal. to the sum of the squared elements of at once that tr (AA / ) = ssq(A). A ,we have Thus (2.4) may be expressed more succinctly as Var(Y , QY) = 2 ' )22 ,/)2 E ssq (X.QX. ~.~. + 4 E ssq ( ~OXoQX. ~ .• .. ~ J ~J. ~ ~ ~J Clearly, if Y/QY is ~o-invariant, which implies contains no elements of quadratic form in ~ 0, (2.5) X;Q = , f3Jo . Thus the variance of any ~O-invariant Y is Var(y/QY) =2 E . . ~J ssq(X~QX.)~~~~ • ~ J ~ (2.6) J Harville also considers a more general type of invariance in regard to the quadratic forms arising from variance component estimation in the one-way classification. What he terms "Q'-invariance" (which implies that the variance of the estimator for the within component contains no terms involving the between component) is expanded here to include invariance with respect to any random or nonrandom effect when estimating the components associated with (2.1). DefiDi tion 2.1: is said to be For any Y defined by (2.1), the quadratic form ~.-invariant ~ m) provided the variance of i (for a fixed value of Y'QY i between does not contain arw = 0 and does not contain any terms involving 2 ~. l a.nd ~o i 0 y' QY and terms if 1= 0 • -e 7 Theorem 2.1: A necessary and sufficient condition for the quadratic , form to be Y QY ~.-invariant l in any non-trivial si tuation is that X~Q ;::: 0 • l Froof: Sufficiency is immediately seen upon examination of (2.5). Recalling that Xm = IN establishes necessity also on inspection. The following lemmas although obvious are presented to show to what extent additional invariance restrictions may be placed on quadratic forms and still maintain unbiasedness when estimating 2 cr. l (i ;::: 1, ... , m) ~.-invariant if J Lemma 2.2: cr.2 l (i = J n. X n for some l l ~uadratic No l for any Denni tion 2.2: l matrix j then it is L. ~.-invariant estimator exists for ~i-invariant estimator exists for crj l unbiased n. X n. l J The quadratic form provided that it is ~.-invariant J 2 L. matrix , Y QY invariant quadratic unbiased estimator of the set ~.-invariant, 1, ... , m) • X. = X.L J X. = X.L is Y QY No quadratic unbiased Lemma 2.3: if I If' a quadratic form Lemma 2.1: is said to be a maximally2 cr. for as many l (i = 1, ... , m) j's as possible among j;::: 0, 1, .•• , m and still maintains its unbiasedness property. By using a maximally-invariant quadratic unbiased estimator, one insures that the Itgoodnes s It of the estimator depends on the vB.lue of the parameter being estimated and on as few of the other parameter ·e 8 values as possible. This property of maximally-invariant estimators is particularly applicable when little or no ~ priori k..l1.owledge is available concerning the relative magni tudes of the variance components being estimated. 2.4 Locall¥ Best Quadratic Unbiased Estimators When the normality of Y is assumed) Rao's [1971a] MINQUE and Rao's [1971bJ MIVQUE estimators coincide and provide locally best invariant 1 ~lE's m ~ 2 K.~. ) where the K. 's are k..l1.own constants. i=l l l l In the context of (2.1) Rao [1972J proves that for any positive definite matrix ditions that for ~o- W) the minimum of , XOQ = 0 and tr QWQW , (i = I) ... ) m) is obtained tr X. QX. = K. l l subject to the con- l when m Q= ~ 0l' RX.X~R ) i=l l l where (2.8) R 0' = and (° 1 ) °2) ••• ) Om) is determined from the equation . . )th ( l)J - where element of 8o = K 8 is ssq(X~RX.) • l J A quadratic form) y' QY) where m E(Y'QY) = !: i=l 1 Q is defined by (2.7 L satisfies 2 K.~. l l ) QUE is henceforth used to denote quadratic unbiased estimator. -e 9 . ( ) = 2 Var Y 'QY and tr QWQW ' )22 Z ssq (X.QX. rr.rr. 1 J 1 J lJ . . 2 tr QVQV , is a minimum subject to the unbiasedness and invariance restrictions. When V the minimum variance t30-invariant W is obtained. When reliable §: ("2 priori estimates rrl' 2 QUE of ZK. rr. ... , ;i) are available for the components being estimated, then the 1 m 1 W matrix of (2.8) may be computed as m W = Z i=l X.X.,'"rr.2 1 1 1 Furthermore, since the restrictive minimization of equivalent to the restrictive minimization of for any is Z p.p. ssq(X~QX.) where .. lJ (i = 1, ..• , m) tr QVQV 1 J 1 J k = 1, 2, •.. , m , the following W matrix may be used in (2.8) to obtain a minimum variance 13 0 invariant QUE of 2 ZK. rr. 1 1 m W= Thus the minimum variance Z i=l X.X~ p. 1 1 1 t30-invariant QUE of 2 ZK. rr. 1 1 may be realized if the ratios of all variance components to a common variance component are known. Rao [1971bJ presents necessary conditions for obtaining a minimum variance t3 -invariant QUE irrespective of the values of the variance 0 components being estimated. required of the X.1 However, due to the restrictive conditions (i = 1, ... , m) matrices, it would seem that few of the generally unbalanced designs met in practice could qualify. Further inspection of these conditions may lead to specially constructed unbalanced designs (perhaps nested) for which minimum variance t30-invariant. QUE's are possible. 10 -e Rao [l972] SUgb2:"~S tLLLt when no ~ priori knO"l'lledge is available for the components being es timo.'cecl Ghi.;:.t the W matrix of (2.8) be computed as m W == Using this L: i==l X.X~ l l W matrix corresponds to assigning equal ~ priori weights to the unknown variance components, and results in the minimization of L: s s q (X ~ QX .) • ij l J Due to the lack of additional invariance constraints, the variance of any t30-invariant QUE in the unbalanced case will, in general, have all terms in the variance expression (2.6) greater than zero. Hence, the "goodness" of any estimator will depend on the actual values of all of the components being estimated. In the event that some components are large relative to others, use of (2.9) as Rao suggests could lead to estimators with possibly undesirable variances. si tuations where QUE's ~ priori knowledge is not would be less riskY. Rao's extended to include t3.-invariant l maximally.,.invariant 2.5 QUE' oS availabl~ maximally-invariant t30~invariant QUE's Perhaps in QUE's are easily (i == l, .•• , m) and thus • Locally Best Quadratic Unbiased Estimators With Additional Invariance Restrictions Locally best t30-invariant invariant QUE's for one or more 13·l which are in addition 13.l (i == l, ... , m) 13Q' denote that collection of 13.l 's here. Let which t3.-invariance is sought. l QUE's are considered for (i == 0, ... , m) In the context of this section t3 Q' -e 1.1 13 is assumed to contain at least f3Q' can not contain (i m) 1, ..• , Obviously, based on Lemma 2.3, . 0 Also denote by 13 13m • not contained in I that collection of The term f3 13Ol Ol 13.l 's -invaria..rlt is used to denote f3.-invariance with respect to all f3. 's l contained in l 130' Furthermore, let 130' 13 and and Oln represent the number of elements in In respectively, (Q'n + In = m+l). 1 by X 11 in , 13 • I X Let ,,2 cr , ... , , ... , Y components cr 2 l of 13 Y Ol ,,2 cr be In cr 2 Yn §: Q'n ,and denote priori estimates of the variance respectively, associated with the elements • The problem, then, is one of finding 13 -invariant QUE's 0' 2 K. cr. , l (where X the incidence matrices associated with the elements In 11 0'1 13 the incidence matrices associated with the e.1ements in ... , x , ... , Denote by KY.' ••. , KY 1 for (2.10) l are a set of predetermined values) for which n ssq(X I QX ) "cr2 ,,2 cr y. y. y. y. J l l (2.11) J is a minimum. In light of Lemma 2.3, (2.10) is not estimable if f3 -invariance 0' implies invariance with respect to any of the elements of which the corresponding f3 y for K value in (2.10) is non-zero. By constructing the matrix In W I . 1:: l=y 1 ,,,2 X.X.cr. , l l l (2.12) -e 12 (2.11) may be rewritten as tr QW QW )' )' and by letting the solution to t3 -invariant QUE's ex is apparent by applying Rao's basic MINQUE theorem from the previous section. t3 -invariant QUE minimized is Y'QY ex To be specific, the for (?lO),if it exists,for which (2.11) is where Q is determined as follows: )'n Q == . l L: 5.RX.X~R, l l l ==)' 1 where and 0 I == (5 , K where , ... , )'1 == is determined from the equation S5 == . . )th ( l,J - S is (KY ' and the element of K 1 ssq(X' RX )' . l )'. ). J The above results imply that if t3. -invariant l more values of QUE's for one or i == 1, ... , m are sought in addition to t30-invariant QUE's, then one has only to include those t3.l 's (for which t3.l invariance is sought) with the set ~O of fixed effects and proceed with Rao's MINQUE estimation procedure as if all the random effects inclUded with t3 0 were fixed. result of the above. The following theorem is an interesting 13 The unique minimum variance Theorem 2. 2: ~aximally-invariant QUE for ~2m in model (2.1) if rank (X) < N is 2 0-m where = y'(I - X(X'xrX')Y/[N - rank(X)], X Proof: The variance of a maximally-invariant y' QY , for QUE, ~ 2 m is SSq(Q)~4m = 2 which is minimized whenever value of 4 tr QIQI QIQI~4 , m is a minimum, regardless of the Thus, to find a maximally-invariant ~ m is necessary to find tr Q From (2.13), 2 tr 1, for QUE ~ 2 m , it Q such that: , X Q 0, and Q = 0lRR , where tr QIQI is a minimum R = I - x(X'X)-X' and 01 is determined from the equation ssq(R)ol = 1 • Since the idempotent matrix rank(X) R is unique, and ssq(R) = tr (R) = N - the theorem is proven. 2.6 A Suggestion For Obtaining Invariant Quadratic Unbiased Estimators ~ a-invariant value of i QUE's between 1 which are also and ~.-invariant 1 for some fixed m have the property that ssg (X~QX,) = 0 J l since ~.-invariance l for all implies j X~Q 0, 1, •.. , m , If (2.7) is computed with a O. l (2.14) matrix W= m ' ,E ,,,,2 X.X. cr. 1=1 where ",2 is an cr. l ~ l l l ,E (2.15) 2 priori estimate of ij , for cr. l i 1 )'" 2t- 2 ssg ( X.QX. cr,cr. l J (2.16 ) J l is minimized subject to the necessary constraints. "practical" . ~. correspon d lng l 1, ..• , m , then It would seem that -invariance could be achieved by replacing the ,,2 cr. l in (2.15) by a number whose magnitude is sufficiently larger than that of any of the is minimized, the coefficients, large values of ~~IS. J Hence when (2.16) ssg(X!Q,X.), associated with the l J will be forced to be relatively small. 15 ·e 3. COMPUTING LOCALLY BEST QUADRATIC UNBIASED ESTIMATORS 3.1 Forming The Matrices Needed (2.7) Although equation ~O-invariant provides the theoretical basis for (2.1), QUE's in models described by a more computationally oriented procedure is given here. Suppose that denoted by ~ priori estimates of r , .•. , r are available. m l Q wi~l 2 v '"'"m (i :=: 1, .•. , m) , Then upon defining r.X.X. ~ i'=l (2.7) and by applying equation v '"'"l2./ , m w matrix Pi -- l l (3.1) l W to this be generated such for some set of K.'s, a l that~ E(y'QY) :=: ~ K.O"~ , ill Var(y'QY) :=: 2 ' ) 0".0". 2 2 ssq (X.Q;X. ~ . . J l lJ l J , and ~ ssq(X~QX.)r.r. ij l J l J will be minimized sUbject to the conditions tr X~ Q;X. :=: K. l l X'Q o o and (i:=: 1, ... , m) • l Thus, values of r. "close" to l will yield an estimator whose variance is "close" to that of the minimum variance QUE of 2 ~ K. 0". ill ~O-invariant 16 -e Using a suitably chosen positive definite matrix W as is given in (3.1), define the following matrices: x~w-:sco A X/W-:SC 0 , = X/W-:SC '-- (3.2) X/W-:SC 0 where o B = (3.3) I and the matrix product: - r-- , BA = X/RX where R = W- l X/RY - W-~o(X~W-~orxbw-l • Since the submatrix I X/RX = X/RX o X/RX of I X RX 1 1 X RX 1 2 X;RX l X .. X/RX ~ m 1 / P2 (3.4) can be partitioned as: ·.. · .. ·· I · X RX m 2 · .. 1 - X1RXm X/RX 2 m ·· · X/RX m m - -(3.4) -e 17 the quanti ties (2.7), equation 8. . lJ = s s q (X l~ RX J.) , (i, j = 1, ... , m) , needed in can be computed directly from (3.4). Q matrix defined by (2.7) is: Also note that the hence m = y' QY o.Y'RX.X~RY 1:: l i=l m ssq (X~RY) sUbmatrix X'RY l 8i ssq(X~RY) • 1:: i=l Thus the element p l (3.5) are also available from (3.4) since the l may be partitioned as: X'RY 2 X'RY = X'RY m The Q. 's l that of (3.5) are computed from the equation -1 8 is non-singular 0 = 8 .•• , m) the elements of K will be: 80 = K and when estimating K • 2 0". l Provided (i := 1, if i := j otherwise Thus, since 8 and 8- 1 (if it eXists) are s~etric, the vector associated with the estimation of 8- 1 • Hence by defining ~, := O"~l will be the [&i, &~, ... , &;J and jth 0 row of 18 T I = [SSq(X~RY), SSq(X;RY), •• " ssq(X~RY)J , (3.6) However, (3.6) may also be derived by equating the T values to their expected values since: E(T. ) 1. m l: j=l 2 ssq(X~RX. )O"j • J 1 Hence E(T) S l: where Since the solution to the system of equations T A = S l: is invariant under linear row operations, provided any set of linear combinations of the equated to its expected value and of full rank. 1. l %Y, i is of full rank, LIT, can be "E be solved for, provided Thus by forming the quadratic forms: y where T. 's , say S = 1, 2, ••. , m , LIS is -e 19 Q1.' := RX,X~R 1. 1. then equating each quadratic form to its expected value and solving, t3 0-invariant for each QUE's variance t30-invariant 3.2 QUE's 2 0-, 1. are obtained which are minimum provided Computing The Although the matrix r i := Pi (i:= 1, •.. , m) . A Matrix A of (3.2) can be computed directly, an alternate method will be presented which in general will require less co~puter storage. Since the matrix W is assumed to be positive definite, there exists a non-singular matrix i would be p:= CA vectors of where the P such that One such W:= pp' P C matrix represents the characteristic W stored columnwise, and Ai represents a diagonal matrix, whose diagonal elements are the square roots of the characteristic W. roots of Unfortunately computer routines require both the upper triangular portion of the W matrix and the C matrix which is N X N to be resident in core. An alt~rnate method for deriving triangular portion of P which requires only the upper W to be in core can be used. Since the Forward Doolittle method as described by Rohde [1964J factors into two triangular matrices, the matrix and Ad Ed ~ matrix, say matrix, such that where Ad is' an upper triangular matrix and is a diagonal matrix with diagonal elements := (DAd)' , where := 1 ~ 1.1. Hence D W, ·e 20 w and by letting P :;;: A/Di , d the W matrix may be expressed as W pp ' (A computational algorithm for forming p' is given in Appendix 7.1). Hence Since the p' described above is an upper triangular matrix, it may be inverted in place using a slightly modified version of the "sweep procedure'outlined by Schatzoff et al. [1968J as described in Appendix 7·2. Thus the A matrix may be computed ~fficiently via the following: STEP 1: Form the upper triangle portion of the W matrix. STEP 2: Use row operations to convert it in place to the Ad matrix of the Forward Doolittle. e STEP 3: Divide each row of the resultant matrix by the square root of the dia,gonal element of' that row. -e 21 STEP 4: Invert the resultant matrix in place using the modified sweep procedure outlined in Appendix 7.2 and name the inverse matrix l' (hence W- l = 1'1 ). Zo = LX O ' STEP 5: Compute STEP 6: Compute the 2; = LX , and G = 1Y A matrix of (3.2) using the matrices of STEP 5 above as follows: ZIG o , A Z' Z0 ___ Z/Z (3.8) Z'G - however, only those elements of A for which i ~ need be j stored. 3.3 Computing The BA Matrix Once the A matrix has been computed using (3.2) or (3.8), by applying Rohde's modification of the Doolittle procedure (to produce a generalized inversion routine) to the sweep routine outlined by Schatzoff (see Appendix nO 7.3), columns to produce the the A matrix may be swept on the first BA matrix of (3.4) in place •. Once this is accomplished, all the elements needed to estimate all variance components are at hand and equation (3.6) may be employed. 3.4 Comvuting Other Invariant Estimators i From the results of section 2.5, locally best which are in addition locally best values of ~.~invariant ~ ~O-invariant QUE's for one or more i = 1, ... , m, may be achieved by augmenting the matrix with the X. ~ matrices associated with those QUE's ~.'s ~ X o for which 22 invariance is sought, and by redefining the estimated variancecovariance matrix as in (2.12). Hence, the preceding techniques may be used to compute estimators which are locally best QUE's in addition to being locally best ~O-invariapt Once locally best ~O-invariant QUE's are computed using (3.6), the associated with (30' achieved. QUE's ~. 1 (3.3) 2 rri (i QUE's. = 1, ••. , m) could be swept on 's as if they were columns The remaining submatrices in could then be used to compute new invariant for BA matrix of the columns associated with one of the ~.-invariant 1 X/RX s and T matrices, and I X RY oDd ~. 1 - for the remaining variance components could be 23 -e 4. QUADRATIC UNBIASED ESTIMATORS FOR THE ONE-WAY CLASSIFICATION 4.1 Foreword The preceding sections outline the theory and the computational procedures for obtaining locally best ~O-invariant variance components associated with model (2.1). the concept of values of ~O-invariance tq include Section 2.5 extends ~i-invariance for one or more i = 1, ••• , m , which in turn leads to the concept of maximally-invariant estimators. heavily restrict the form of general (except for ~ QUE's for the 2 ). m 0- Although maximally-invariant QUE's Q, uniqueness ~an not be claimed in Thus, in the "generally" unbalanced case priori estimates of the variance components (for which invariance is not sought) must be supplied to completely specify the estimator for a particular component. When ~ priori estimates are not available, are there other relatively "safe" prior estimates, based on the design characteristics, which could be used to p,revent the estimator from having a variance much larger than that of the "best" estimator were the true variance components known? In seeking a partial answer to this question, we take the approach suggested by Harville and work with a simplified model, that of the unbalanced one-way classi- fication. As stated by Harville, "virtually every problem that arises in constructing a complete bOdy of theory for the estimation of the variance components associated with a comPlex possiblyunbalanced random or mixed model is also encountered, though in simpler form, in carrying out that process for the possibly-unbalanced one-way 24 random lUodel". By extending our knowledge of this model, "we effectively place an 'upper-bound' on what we can hope to achieve in the way of theoretical results for more complicated models". 4.2 ~he Mathematical Model The one-way classification model is represented here as: (4.1) where •.• , Yal' •.• , Yan ) a X is an o ~ NX 1 vector of l's, is a fixed but unknown constant, Xl is an NX a matrix whose for l's in elements R jth column is zero except thru 'R l 2 where + ••• + n .. and J a' = (aI' ... , aa) where a is assumed to be distributed 2 Normal (0, laO'Q') , E' = ) where E is , ..• , Eal , ••. , E an lnr a 2 assumed to be distributed Normal (0, IN~e) , a.l'ld a.1 (Ell' ..• , and E.. J.J E for all correlated. i and j are assumed to be un- 25 ·e From the above, y ,.., Normal (XolJ., V) , where 4.3 Invariant Quadratic Unbiased Estimators For The Variance Components Although Harville provides locally best IJ.-invariant cr 2 e The variance of any J,L-invariant and l for through use of Lagrangian functions, a simplier derivation is given here by applying equation (K QUE's QUE, y' QY ,for l<lcr; + ~cr; are predetermined constants) is: K 2 Var(y'QY) (2.7). =2 4 2 2 4 ssq(X 'QX1)cr + 4 ssq(X 'Q)cr cr + 2 ssq(Q)cr , l a a e e l where Hence, infinitely many IJ.-invariant QUE's for generated by choosing the matrix tr Qr = ~ Q r ~ ~cr: priori estimates, r's, of such that: X'Q or + p K2cr~ can be and determining "'·0, , and minimizing (4.2) -e 26 When 11.0-: r = p , the minimum variance + ~C1; we have from (2.7) that the y'Q Y minimized (4.2) is 01 QUE for is obtained. Since (4.2) is equivalent to and ~-invariant and 02 r ' tr Qr Wr Qr Wr ~-invariant QUE where of where: are determined from the equation: ssq(X~RX1) ssq(X~R) Kl 61 (4.4) ssq(X~R) Since -1 W ssq(R) 6 K 2 2 can be expressed in closed form, (4.3 ) simplifies to: C11J 11 C12J 12 C1aJ la C12J 21 C J 22 22 C J 2a 2a + 10 2 J Qr = C J 1a ai where each e C.. ~J J .. ~J C2aJ 2a represents an (4.5) C J aa aa ni X n. J matrix of l'sJ and each .. + h.. +l(i=j)] + 20 t .. + °2mij ° 1[(ni + n.J )t ~J 2 lJ lJ . -e 27 The quap.ti ty ~=J =j and zero otherwise. and m .. are computed as follows: gi = r/(l value of one if t .. , ~J h.. , ~J represents the indicator function, which has a l-(. .) i iJ The remaining quantities Let + rn ) i for (4.6) i = 1, ..• , a , and G = -lIeN - ~. n~g.) ~ ~ ~ Then and The elements of (4.4) can be expressed as: ssg(R) =N+ 2 ~ n.n.t .. + 2 ~ nkt kk ' ij ~ J ~J k 122 2 Z n.n.t' j + 2 Z nkt kk , ij ~ J ~ k ssg(X1R) = N + = Z n k2 k + ~ ij 222 n.n.t .. + 2 ~ J ~J 3 and ~ nktk~ k ~ 28 ·e Although the above equations could possibly be simplified, they do serve the purpose of reducing the amount of computer storage required to study I-L-invariant QUE I s of lS.cr~ + K2 cr: which minimize (4.2). Examination of (4.5) leads to the fact that the estimator l is a function of the sufficient statistics 2 Z y... . . lJ l,J the class sums) and ••• , Y Y'Q Y r a' ,i.e., -- Furthermore (4.5) meets the necessary condi tions specified by Harville [1969a] for "quadmissibili tylf since if For any n. l Q r = n~ l and n. J = n , j • matrix computed using (4.5), and = 4 ,2 2 2 ssq(Xl'Qr Xl)cra + 4 ssq(X1Qr)cr cr ae 4 + 2 ssq(Q)cr , r e where = 4 Z n. (n.C .. + 02) + 2 i l l II Z i,j>i 2 n.n.(n. + n.)C .. , l J l J lJ and 2 ssq(Qr ) = 22 2 Z [no (C .. + 02) 2 +(n. - n. )C .. J + 4 i l II l 1 See Hultquist and Graybill [1965J. l II 2 .. Z n.n.C l J lJ i,j>i 29 -e The only restriction that need be placed on the value of -lin. that it not be equal to for any ~ i = ... , 1, r is a ; in this case (4.6) becomes infinite for some value of i . 4.4 Relative Efftciencies Having stated the model and provided a technique for infini tely many }.1-invariant QUE I S for gener~ting 2 2 Kl (J"a + !<2(J"e ' we are now in a position to study the behavior of the variances of estimators generated by using (4.5) with a variety of values for ~ ~ 0 , and to compare them with the variance of the. minimum variance }.1-invariant QUE. Throughout this section Q r will denote the matrix generated by (4.5) for a specified value of r of Kl K2' and (4.5) with The matrix ~ Q 0 denotes the matrix generated by p =p r KJ.. and For any values of K 2 minimum variance }.1-invariant QUE for for any for any r ~ and for some specified values 0 r~lative to (not both zero), KJ..(J"; r F 0 • + K2 0": . Y'Q Y is the P The The efficiency of y'Q Y r Y'QpY denoted by Since little can be shown theoretically about Eff(Qr IQ p ) due to the algebraic complexity of the problem, the current author employed equation (4.5) in a computer program to examine Eff(Qr IQp ) over a range of values of r and p classification designs. Tables for a number of different one-way 4.1, 4.2, and 4.3 are a sample of the tables produces in studying the efficienty of the }.1-invariant QUE for 2 (J"a (K l = 1, K2 = 0). Tables 4.4, 4.5, and 4.6 are a sample of the Table 4.1 r=O r=.25 r=l r=5 r=10 r=100 r=1000 r=10000 1 ANOVA e e e Eff (Qr IQp ) when estimating cr; in a one-way classification model with cell frequencies: 3, 5, 59, 20, 50, 21, and 89 p=O p=.25 p=l p=5 p=10 p=100 p=1000 p=10000 1.000000 0.150487 0.053123 0.034322 0.032262 0.030465 0.030288 0.030271 0.587925 0.545024 1.000000 0.898088 0.788522 0·770279 0·752912 0.751126 0.750947 0.807093 0.441949 0·933258 1.000000 0·981cxJ7 0.974931 0·968404 0·967695 0.967624 0.671921 0.399638 0.874939 0.983854 1.000000 0.999613 0·998537 0·998388 0·998373 0.611748 0.393582 0.865512 0·979049 0·999621 1. 000000 0·999662 0·999589 0·999581 0.602986 0.387957 0.856543 0·974cxJ9 0·998596 0·999668 1. 000000 0.387385 0.855619 0·973535 0·998456 0.999598 0·999997 1. 000000 1.00000- 0.387328 0.855527 0·973481 0·998441 0·999590 0·999996 1.000001.000000 0·593903 0·999997 0·999996 0.594819 0.593987 1 The efficiency of the standard ANOVA estimator (described in section 4.6) is included here for comparative purposes \>I o e e Table 4.2 Eff(Qr IQ) when estimating rra2 in a one-way classification model with cell frequencies: p 22, 52, 33, 88, 68, 48, and 25 p=O r=O r=.25 r=l r=5 r=10 r=100 r=1000 r=10000 ;;NaVAl e 1. 000000 0.559071 0·513653 0·500592 0.498931 0.497430 0.497279 0.497264 0.844936 p=.25 p=l p=5 p=lO p=lOO p=1000 p=10000 0.655821 1.000000 0.996715 0.994470 0.994140 0·993833 0.993802 0.620161 0·996790 1.000000 0·999713 0.999635 0·999555 0·999547 0.999546 0.859781 0.609706 0.608368 0·994316 0·999637 0·999995 1.000000 0.999996 0·999995 0·999995 0.848586 0.607159 0.607037 0·993992 0·999551 0·999981 0·999995 1.00000 1.000000 1. 000000.847303 0.607025 0·993989 0·999550 0·999981 0·999995 1.000001.000001.000000 0.841291 0·993799 0.891645 0.994632 0·999715 1. 000000 0·999995 0.999983 0.999981 0.999981 0.849871 0~994022 0·999559 0.999983 0·999996 1. 000000 1.000001. 000000.847420 1 The effici€ncy of the standard ANeVA estimator (described in section 4.6) is included here for comparative purposes \.>J i-' e e Table 4.3 r=O r=.25 r=l r=5 r=10 r=100 r=1000 r=10000 PJifOVA1 e when estimating ~2a in a one-way classification model with cell frequencies: Eff(Qr IQ) p 1, 33, 94, 78, 1, 64, 91, 69, 72, 1, 24, and 42 p=O p=.25 p=l p=5 p=10 p=100 p=1000 p=10000 1.000000 0.184756 0.01ll97 0.002280 0.001775 0.001401 0.001367 0.001364 0.646481 0·743460 1.000000 0·766674 0.338290 0.280231 0.231693 0.227077 0.226618 0.906109 0.662315 0·934801 1.000000 0.863808 0.805017 0·739401 0·732190 0·731462 o~ 816273 0.571334 0.815404 0.550680 0·786971 0.898059 0·995820 1.000000 0.528928 0.756779 0.867824 0·983743 0·995623 1.000000 0.526564 0·753486 0.864453 0·982032 0.994662 0·526326 0·753154 0.864113 0·981856 0·994560 0·999937 0·999999 1.000000 0.650962 0·925139 1. 000000 0·995427 0·980770 0.978587 0.978361 0·706138 0·995258 0·994171 0·994055 0.680848 0·999947 0·999936 0.654158 0·999948 1.000000 0·999999 0.651255 1 The efficiency of the standard ANOVA estimator (described in section 4.6) is included here for comparative purposes 'vi I\) e e T-able 4.4 e 2 Eff(Q.-r IQ) when estimating rre in a one-way classification mode~ with cell frequencies: p 3, 5, 59, 20, 50, 21, and 89 p=O p=.25 p=l p=5 p=10 p=100 p=1000 ------ .p=l-DOOO r=O 1.000000 0.140769 0.000017 0.000000 0.000000 0.992288 0.006756 0.838361 0.001706 r=.25 0·700228 1.000000 0.569698 0.000001 r=l 0·990311 0·999020 0.000136 0.013387 0.434284 . 0.007642 0.000077 r=5 r=10 0·989969 0·996338 0·730918 0.026453 0·999755 1.000000 0.286479 0·999814 0·999995 0·999990 0.975695 r=100 0.989955 0.989949 0·998695 0·998681 0.998676 r=1000 0.989949 0·998676 0.999814 0·999990 0·999997 1.00000- 0·999732 1.00000- r=10000 1 ANOVA 0.989949 0.998676 0·9998J..4 0·999990 0·999997 1.00000- 0·999997 1. 000000 1. 00000- 0·989949 0.998676 0·999814 0·999990 0·999997 1.00000- 1.00000- ........-. _. '- 0·992585 1.000000 0·999833 0.999819 "#. ._.--... - 0·996873 1. 000000 0·987058 0·999979 1. 000000 0·999997 1 1.000000 1.00000- 1 The efficiency of the standard ANOVA estimator (described in section 4.6) is included here for comparative purposes VJ VJ Table 4.5 -- e e Eff(Q IQ) when estimating ~2 r p e 22, 52, 33, 88, 68, 48, and 25 p=O in a one-way classification model with cell frequencies: p=.25 p=l p=5 p=lO p==lOO p=lOOO p=lOOOO r=O l. 000000 0·707l0l 0.l42656 0.006806 0.00l7l7 0.0000l7 0.000000 0.000000 r=.25 0.9972';f8 l.OOOOOO 0·99ll0l 0·965369 0.2l8622 0.002792 0.000028 r=l 0·997227 0.999976 0·999675 l. 000000 0·999B09 0·980934 0.339785 0.005l2l r=5 0·997223 0·999972 0·999998 0·999955 l. 000000 0·999966 0·996596 0.745375 r=lO 0·997223 0·999972 0·997223 0·999972 l.OOOOO - l. 00000 0·999998 l. 000000 0·999784 l.OOOOO - 0·978849 r=lOO 0·999998 0.999998 r=lOOO 0·997223 0·999972 0.999998 r=lOOOO l ANOVA 0·997223 0·999972 0·999998 l.OOOOO - 0·997223 0·999972 0.999998 l.OOOOO l. 00000 - l. 00000 - l.OOOOO - l.OOOOO l.OOOOOO l.OOOOO l. 00000 - l. 00000 l.OOOOOO - l. 00000- l. 00000 l. 00000 l. 00000 - - 0·999998 l.OOOOO l. 000000 l. 00000- l The efficiency of the standard ANOVA estimator (described in section 4.6) is included here for comparative purposes \.>J +" e e e Table 4.6 Eff(Qr~IQ) p when estimating ~2e in a one-way classification model with cell frequencies: 1, 33, 94, 78, 1, 64, 91, 69, 72, 1, 24, and 42 p=O p=.25 p=l p=5 p=lO p=lOO p=lOOO p=lOOOO 0.201057 0.-010270 0.002597 0.000026 0.000000 0.000000 0.997504 0·785052 l. 000000 0.995835 0·908862 0·728468 0.028563 0.000297 0.000003 r=l 0.995061 0·998645 l. 000000 0.962757 0.216443 0.002787 0.000028 r=5 r=lO 0.993360 0.997080 0.247723 0.003286 0.996942 0·999949 0·999822 l. 000000 0·970352 0·993222 0·999123 0.998990 0·990550 1.000000 -0.802567 0.039101 r=lOO 0.993166 0.996887 0·998935 0·999902 0·999972 r=lOOO 0·993166 0.993166 0.996886 0·998935 0·999901 0·999972 0·997570 l. 000000 1.00000- 0·999969 1.000000 0·996867 1.00000- 0·996886 0·998935 0·999901 0·999972 1.00000- 1.00000- 1.000000 0·993166 0·996886 0·998935 0·999901 0·999972 l. 00000- 1.00000- 1.00000- r=O 1.000000 r=.25 r=lOOOO 1 ANOVA I 1 The efficiency of the standard ANOVA estimator (described in section 4.6) is included here for com~arative ~ur~oses \.).J \Jl 36 tables produced in studying the efficiency of the ~-invariant QUE for 2 ere (Kl :::: 0, ~ :::: 1) • 4.1 thru 4.6 seems to indicate the Close examination of Tables follow:tng: when estimating approaches P or 2 er e using equation (fixed) from above or below, increases to a value of one. approaches 2 er a r Eff(Q IQ) Furthermore, it appears that as Eff(Q IQ) r P p (fixed). In tables not shown, the = n2 to one any time 4.5 :;::: ••• == n P Eff(~IQp) ere' seems to approach an asymptotic value rather quickly as beyond r monotonically 2 Also, when estimating as monotonically P r (fixed) from above or below, increases to a value of one. (4.5), r Eff(Qr IQp ) increases was equal a Choosing Reliable Estimates With i Limited Prior Knowledge If a reliable estimate is avaLI.able for estima~e 2 er a anq. in er (4.5) 2 e p, then use of that should yield fairly efficient estimators for both due to the seemingly monotonic prope~ties In many situations it might be possible to bracket upper and lower bounds. true value of p Eff(Qr IQp ) • with feasible In other words, if we are confident that the lies somewhere between, say, choosing a particular value of (4.5) p of r between Po and and Pl ' then to use in would seem the most logical choice. By observing Tables 4.1 thru 4.6, one can see that tbe line associated with any choice of r has its smallest efficiency either in the first or last position of the line. This smallest efficiency -e 37 represents the "worst" one can do if he uses that particular the true value of P bounds of the table. between and P r and in fact lies somewhere within the upper and lower If we are certain that the true value of l ' the~for any value of r P +les we use in (4.5), the worst efficiency we could possibly have would be the smaller of either Eff(~IQ ) or Eff(Q IQ ). Conceptually, we could build ~ list Po r Pl containing all possible r values along with the worst possible efficiencies associated with each of them assuming and of r , we could pick the r lies between To minimize the risk we are taking in choosing a value r value aspociated with the largest efficiency in this list (of worst efficiencies). of P By using this value in (4.5) we could provide not only the variance component estimates based on this r value, but also a "guaranteed" efficiency level for the estimators, which would merely be the worst possible efficiency associated with that value of 4.1 thru 4.6, one can see that the On closer examination of Tables search area for the "best" r and Pl. r* then r* If we denote by may be restricted to values between the value chosen is equal to the value of F(r) = min[Eff(Q IQ r is a maximum over the range r. Po $ r Po ), Po by this technique, for which Eff(Q IQ )J r P r S P . l (4.7) 38 -e The following graph (Figure 4.1) illustrates the above points. Efficiency 1 r Figure 4.1 r * Typical efficiencies when estimating (J 2 or a (J 2 e Although the above curves are merely representations of the efficiencies, the following point can be made: has a value of 1 1I11.o:;;J. approaches P l ;) o from below and since F(r) Eff(Q IQ r and monotonica.1J.,y decreases as Eff(Q IQ and monotonically decreases as the function Since r r P l Po ) r ) has a value of 1 when approaches Po ftom above, of (4.7) is represented by the portion of the two curves closest to the r axis. As can be seen F(r) achieves a maximum at the point of intersection of the two curves, and it is at this intersection that r* is realized. Since the theoretical formulas for these curves are intractable, can be used for determining the following algorithm r* Algorithm 4.1: STEP 0: Assign the desired values for Kl' K, 2 and set EO = desired accuracy of F(r*) (~.~., the algorithm terminates -e 39 when a value of IEff(Q IQ r Set STEP 2: Set STEP 3: Compute 4: Po is found sUGh that ) ~ Eff(Q IQ )\ ~ EO) . r Pl and STEP 1: STEP r r r = L + (r H - r )/2 • L Qr using equation Set F STEP 5 : If IF\ STEP 6: If F < 0 , go to STEP = (4.5). Eff(Q IQ ) - Eff(Q IQ ) r Po r Pl ~ EO' set r * = r and terminate, otherwise: 7; otherwise set r L r , and go to = STEP 2 . STEP 7: Set r H =r and go to STEP 2. r* Since the search area for is halved at each iteration in the above algorithm, convergence is quite rapid. presented in sectio~ Several examples are 4.6 which compare the Qr * estimators (r* being generated by algorithm 4.1) with the standard analysis of variance ~2 estimators for a and ~2 e When no knowledge about P is available other than assuming P ~ 0 , then algorithm 4.1 may be applied by setting equal to a pseUdo value for infinity. instances suffice. Of course the smaller the better, especially when estimating 2 ~e A choice of P l ~ 2 a ' one need not be concerned about making since it appears that for any fixed value of p. o=0 and Pi = 1000 would in m2-st can prudently be made However, when estimating P l as small as possible, P, the approaches an asymptotic value quite rapidly as the value of P r Eff(Q IQ ) r P increases beyond In other words, if it were known that p were less 40 than say 100, then any choice of PI ~ 100 , whether it be 100 or 10,000 would not make any appreciable change in the choice of Based on the above discussion, we r* uses the Definition r* * define an estimator which (4.1). of algorithm 4.1: The Relatively Safe Quadratic Unbiased Estimator 2 + Kl~a ( RESQUE ) for where for~ally r 2 K2~e , Y Q *Y , in the one-way classification is is determined by algorithm r 4.1 and Q * is determined by r (4.5) equation 4.6 with r r * Comparison of Invariant Quadratic Unbiased Estimators To The Sta~dard Analysis of Variance Es~imators The standard analy$is of variance estimators for ~ 2 a are: 0-2 = Y'EY , e and 0-a2 y'AY, where a-I N-a I] , and where A is an elements are etc. Both 1/n Y'EY l N X N diagonal matrix whose first , whose next and y'AY n 2 n diagonal elements are l diagonal 1/n 2 are ~-invariant estimators, and the , -e 41 matrices of E (4.5). and A have the same structl,lre as does the In terms of (4.5), the 1 matrix Qr E matrix has: 1 -(~)(-) if N-a n.l i == j if i c.. lJ o otherwise and 6 In terms of (4.5), 2 1 ~ == N-a A matrix has: the K ~ K a (N-l) N-a n. l N j K a otherwise N and (a-i), K 62 a N-a where K N a Since both the A and Qr ' is there a value of the A and the r E matrices? E matrices have the same structure as which could be used in known, setting r:::; co. to generate By Theorem (2.2), the estimator is a maximally-invariant estimator. generated by setting (4.6) Therefore the E matrix may be Hence, if an upper bound on p is + equal to that upper bound will produce a more efficient estimator than the standard analysis of variance estimator 42 -e The corresponding value of not as easily determined. l' associated wi th the matrix However, by making numerous computer runs, the current writer found that a value of harmonic mean of the I' = Iln h n.l 's ) produced a matrix approximately equaled Y'AY Iln h p (where n such that for all cases tested. knowledge is available which indicates that from A is h is the Y'Q Y l' Thus, if prior has a value diffe:!:'ing , then use of that knowledge in algorithm 4.1 should pro- duce a more efficient estimator. To compare further the standard analysis of variance estimators wi th ~-invariant QUE's, the following tables give the worst efficiency achievable provided p l i es between of variance estimators, the based on setting r =1 RESQUE in (4.5). and Pl ' for the analysis estimators, and the estimators -e 43 Table 4.7 Component to be Estimated cr2 a cr 2 e Table 4.8 Smallest obtainable efficiencies when 0 ~ p ~ 10000 of RESQUE, ANOVA and Qro1 estimators for a one-way classification model with cell frequencies: 3, 5, 59, 20, 50, 21, and 89 Smallest RESQUE Efficienoy Smallest ANOVA Efficiency Smallest Qr=l Efficiency .0427 .62780 .58792 .05312 40.2cb9 ·98994 ·98994 .00007 RESQUE r * value Smallest obtainable efficiencies when 1 ~ P ~ 10 of RESQUE, ANOVA and Qr=l estimators for a one-way classification model with cell ftequencies: 21~ and 89 Compo.qent to be Estimated 2 cr a cr 2 e Table 4.9 Component to be Estimated cr 2 a cr2 e 3, 5, 59, 20, 50, Smallest RESQUE Efficiency .Smallest ANOVA Efficiency Smallest Qr=l Efficiency 1·9097 ·99419 .60298 ·97904 3.2500 ·99985 .99981 ·98705 RESQUE r * value Smallest obta~nable efficiencies when O~· P ~ 10000 of RESQUE, ANOVA and Qr=l estimators for a one-way classification model with cell frequencies: 22, 52, 33, 88, 68, 48, and 25 Smallest RESQUE Efficiency Smallest ANOVA Efficiency Smallest Qr=l Efficiency .0222 .84952 .84493 ·51365 16·7847 ·99722 ·99722 .00512 RESQUE r * value -e 44 Table 4.10 Smallest obtainable efficiencies when 1 ~ P ~ 10 of RESQUE, ANOVA and Qr=l estimators for a one-way classification model with cell frequencies: 22, 52, 33, 88, 68, 48, and 25 Component to be Estimated (J" (J" 2 a 2 e Table 4.11 Component to be Estimated (J" (J" 2 a 2 e Table 4.12 Smallest RESQUE Efficiency Smallest ANOVA Efficiency Smallest Qr=l Efficiency 1.8438 ·99990 .84858 .99963 5·5000 ,99999 ·99999 ·99980 RESQUE r * value Smallest obtainable efficiencies when 0 ~ p s 10000 of RESQUE, ANOVA and ~=l estimators far a one-way classification model with cell frequencies: 1 1 33, 94, 78, 1, ?4" 91, 69" 7~,· 1, 24, and 42 Smallest RESQUE Efficiency Smallest ANOVA Efficiency Smallest Qr=l Efficiency .057 4 ·70109 .64648 .01119 82,0923 .99316 ·99316 .00002 RESQUE r * value Smallest obtainable efficiencies when 1 ~ p ~ 10 of RESQUE, ANOVA and ~=l estimators for a one-\\'ay clasi3ification model with cell frequencies: 1, 33, 94, 78, 1, 64, 91, 69, 72, 1, 24, and 42 • Component to be Estimated 2 (J" a 2 (J" e Smallest Qr=l Efficiency RESQUE r * value Effi~iency Smallest ANOVA Efficiency 2.0624 .96166 .68084 .89805 3.6016 ·99925 ·99893 .96275 Smallest RESQUE 5. SUMMARY.AND SUGGESTIONS FOR FUTURE RESEARCH 5.1 Summary Rao's recent works on MINQUE and MIVQUE provide estimators of the variance components in linear models. If normality is assumed, then MINQUE and MIVQUE coincide, and provide locally best quadratic unbiased estimators which are effects in the model. invari~t to translation of the fixed As pointed out in section 2, estimators, which are based on ~ ~o-invariant priori estimates of the components being estimated, have variances which are, in general, functionally dependent on all of the true values of the parameters in the model. The basic weakness of MIVQUE is then, that when ~ priori estimates are not "close" enough to their true values, or when no §: priori estimates are available, the variance of MIVQUE estimators could in many instances be unacceptably large. To help reduce this risk, the concept of in- variance with respect to one or more of the random components was introduced. These additional invariance concepts, remove certain terms from the variance of an estimator in order to protect against situations for which these terms might be large. The concept of maximally-invariant estimators was introduced to denote those estimators for which the variances are as free as possible of terms involving other components in the model. In section 3, techniques are presented for computing MIVQUE estimators, and also for locally best quadratic unbiased estimators which are invariant to other random effects in the model. In that section, Qne can see that MIVQUE is actually equivalent to selecting a set of sums of squares from a weighted regression analysis, to equate 46 them to their expected values, and to solve to get variance component estimates. To provide some insight into the choice of the prior estimates to use when either limited or no i~ MIV~UE, able, section ~ priori knowledge is avail- 4 examines what can be done in this respect when working with the one-way classification. Relatively Safe ~uadratic A new estimator, referred to as the Unbiased Estimator (RES~UE), is developed which in some sense minimizes the risk one takes in choosing the prior estimates of components to use in 5.2 MIV~UE. Suggestions For Future Research One of the problems which arises in weighting matrix which must be inverted. MIV~UE is the size of the Although in section 3 techniques which can be used are presented, these still require in general the inversion of the N X N weighting matrix (where presents the number of observations). N re- Unless techniques can be developed which circumvent this inversion process, rendered an interesting but somewhat useless MIN~UE technique~ will be Along this same line, using an identity matrix as the weighting matrix might prove quite efficient for those designs which are not very unbalanced, such as those arising from experiments which were initially balanced, but for which some observations have been lost. One of the other computational prob~e~s, so far as the size is Gonoerned, is the amount of storage required to compute the portion of (3.4). . 1S I rr2 ,where e X/RX Assuming the last effect in the model being analyzed I is N X N and 2 rr e is the error variance, then 47 the quantities ssq(X~R), ssq(X;R), ••• , ssq(X~_lR), and ssq(R) are needed in the expectation of each sums of squares being computed. Since the " X R, ••• , Xm_1R , , and X1R, 2 R represent the major portion of X'RX matrix, computing them in some other fashion would reduce considerab~ the amount of computer storage required. In addition to the computational problems described above, there remains the problem of what can be done when no available. c~n ~ priori estimates are Hopefully the work presented in section 4 describing what be done in this situation for the one-way classification can be extended to cover the more general linear models discussed in section Although Rao [1971bJ presents the basic theory for obtaining minimum mean square quadratic estimators when the true values of the variance components are known, Rao [1972J suggests that iterative estimation using MINQUE's may provide estimators with interesting properties. These "interesting properties" might include sma.ll mean square errors, obtained through the use of equation (4.5) and the more general techniques of section 3. As aid to future research in this area, and to make available MIVQUE's for experiments of under 250 observations, the present writer has developed a computer procedure which is available in the statistical Analysis System of North Carolina State University at Raleigh. A description of the system and the MIVQUE procedure is given by Service [1972J. 48 -e 6. LIST OF REFERENCES Harville, D. A. [1969aJ. Quadratic unbiased estimation of variance components for the one-way classificatiop. Biometrika 56:313-326. Harville, D. A. [1969b J. Variance com.ponent estimation for unbalanced one-way random classification - a critique. ARL Report 69-1080. Henderson, C. R. components. [1953J. Estimation of variance and covariance Biometrics 9:226-252. Hultquist, R. A. and F. A. Graybill [1965J. Minimal sufficient statistics for the two-way classification mixed model design. JASA 60:182-192. Koch, G. G. [1967J. A general approach to the estimation of components. Technometrics 9:93-118. vari~lce Rao, C. R. [1970J. Estimation of heteroscedastic variances in linear models. JASA 65 :161-172. Rao, C. R. [1971aJ. Estimation of varianc~ and covariance components MINQUE theory. Journal of Multivariate Analysis 1:257-275. Rao, C. R. [1971bJ. Minimum variance quadratic unbiased estimation of variance components. Journal of Multivariate Analysis 1:445-456. Rao, C. R. [1972J. Estimation of variance and covariance components in linear models. JASA 67 :112-115. Read, R. R. [1961J. On quadratic estimates of the interclass variance for unbalanced designs. J.R.S,S. B 23:493-497. Rohde, C. A. [1964J. Contributions to the theory, computation, aDd application of generalized inverses. Ph.D. Thesis, Department of Experimental Statistics, North Carolina State University at Raleigh, Raleigh, North Carolina. Untverisity Microfilms, Ann Arbor, Michigan. Schat~off, M., R. Tsao, and S. Fienberg [1968J. Efficient calculation of all possible regressions. Technometrics 10:769-779. Searle, S. R. [1971J. Topics in variance com.ponents estimation. Biometrics 27:1-76. Service, J. W. [1972J. A User's Guide To The Statistical Analysis System. North Carolina State University Students Supply Stores, Raleigh, North Carolina. -e • Townsend, E.C. [1968J. Unbiased estimators of variance components in simple unbalanced designs. Ph.D. Thesis, Biometrics Unit, Cornell University, Ithaca, New York. University Microfilms, Ann Arbor, Michigan. 50 • 7. APPENDIX 51 7.1 An Algorithm For Factoring A Symmetric Positive • Definite Matrix in Place Denoting by W the symmetric positive definite matrix which is to be factored, only the upper triangular portion need be computed and stored in core. The algorithm presented here computes the P I out- lined in (3.7) and stores it in the same core positions previously occupied by the upper triangular portion of W without using any other storage locations for intermediate results. STEP 0: Compute the N(N+l)/2 elements of Wand denote them by Wi j (i::;: 1, ••. , N), (j ::;: i, ..., N) STEP 1: Set k STEP 2: Set STEP 3: Set k + 1 and set K + 1 and if D and set W • kk := l > N , then go to STEP Wij ::;: Wij - WkiWkj/D k::;: 0 • for each 5. (i::;: l, ••. , N) and (j ::;: i, ... , N) • STEP 4: If k < N, then go to STEP 1. STEP 5: Set i::;: 0 STEP 6: ~et i ::;: i + 1 STEP 7: pet STEP 8: If and Wij ::;: W1' J,/D D=,r;!. '\!wi i for each i < N , then go to STEP labeled W ij are now j ::;: i, ..• , n • 6, otherwise stop; the elements P~ .• 1J The above algorithm stores only the upper triangular portion of · le '1S a 11 zeros. P I sl'nee the 1 ower tr1ang In the event that the matrix W is not of full rank, a singularity check may be inserted folLowing STEP 1 as follows. 52 STEP la: D< If ~ero), 7.2 E l (where E then stop, l > 0 represents a pseudo value for W is singular. An Algorithm For The Inversion Of An Upper Triangular Matrix in Place Assuming that the elements above and inclUding the diagonal of an upper triangular matrix denoted by W.. lJ ••• , N) ~ (i 1, ••. , N), (j = i, have been stored, the following algorithm replaces the elements of -1 W W with the elements of without additional working storage needed, and is a modification to the sweep technique presented by Schatzoff, et~. STEP 0: Set k ::;: 0 STEP 1: Set k ::;: k + 1 STEP 2: Set i ::;: i + 1 set j = k - 1 Set j set W.. lJ STEP 3: STEP 4: If STEP 5: If STEP 6: Set STEP j < ::;: , j + 1 ::;: D ::;: W kk and if i ::;: k set and if j , an,d set , then go to STEP 5; otherwise k , then go to STEP W.. - wikwkj/D lJ N , then go to STEP 3 Set STEP 9: If 4; otherwise . i < k , then go to STEP 2 • W ::;: -Wil!D ik for each i = 1, 7: Set Wkj ::;: Wk/D for each j :: k, STEP 8: i = 0 • W kk ... , k • ... , N • = liD. k < N , then go to STEP 1; otherwise stop; the upper triangular portion of W has been replaced by the upper triangular portion of W -1 • 53 The above algorithm takes advantage of the fact that the i~v8rse of an upper triangular matrix is an upper triangular matrix and hence neither W nor be stored. -1 W needs the lower triangular portion of' In the event that check as was given in Appendix 7.3 to z:-~ros W might be singular, a singularity 7.1 may be inserted following STEP 1. An In Place Generalized Inverse Sweep Algorithm As pointed out by Rohde the abbreviated Doolittle method may be modified to produce a generalized inverse by setting any row of the Ad matrix of the forward Doolittle to zero if its diagonal element goes to zero during the forward Doolittle procedure, and by setting the corresponding column of the B d diagonal element which is set of 1. to the Ad and B d generalized inverses. matrix to zero except for the ApplYing the backward Doolittle matrices thus defined, then produces a This same technique applied to the sweep routine outlined by Schatzoff et al. would imply that if any pivat element goes to zero then the matrix will be considered swept on that pivot ele~ent once the row and column containing that pivot element have been set to zero. Using the modified sweep routine on pivot elements 1, 2, ••• , nO of the matrix A of (3.2) would thus be equivalent to mUltiplying by the matrix B of (3.3) yielding the resultant matrix A BA produ~t of (3.4). Assuming only the elements stored (!.~., (i n c A on and above the diagonal have been = 1, ••. , n r ) and (j = i, ••. , n c )) ,where represent the number of rows and columns respectively of nand .r A, the follo¥ing algorithm performs a sweep on each of the pivot elements I' 1, •.• , nO ' with the result that the elements the corresponding elements of the STE:f 0: Set k = 0 • STEP 1: Set k STEP 2: If = k + 1 D > €l and D (where BA matrix of l STEP 3: 4: Akj := for 0 i =i + set j =i STEP 5: Set j = j + 1 STEP 6: Set B = gik C = -1 i = for 0 ... , nr = k then go to STEP 8·, otherwise and if if j = k , then go to STEP 7. i < k , and , and otherwise A {~ -A if k < j otherwise - BC/D • A.. = A.. lJ lJ j <n , then go to STEP 5· c If STEP 8: i < n , then go to STEP 4. r Set Aik = -AiJD for each i = 1, If Set Set ~ STEP 10: and. 1, ••• , k ; then go to STEP 10. STEP 7: STEP 9: i . kj set ik = k, and if 1 ki set A . Set i = 0 Set STEP j (3.4). is a pseudo value for zero), then go to STEP 3; otherwise set set lJ Akk • = E A.. If Akj = ~/D for each j = k, ... , k ... , n c Akk = lID • k < nO ' then go to STEP 1·, otherwise stop; the matrix of (3.4) has replaced the A matrix of (z. .. ./ • 0) • c:.. BA
© Copyright 2024 Paperzz