• JACKKNIFING MAXIMUM LIKELIHOOD ESTIMATES IN THE CASE OF A SINGLE PARAMETE~ By R. A. Ferg~son and J. G. Fryer D~partment of Biostatistics University Qf North-Carolina at Chapel Rill and University of Exeter, England Institute of Statistics Mimeo Series No. 891 SEPTEMBER 1973 The main purpose of this paper is to illustrate and quantify the effect that jackknifing can have on the lower sampling moments of a maximum likelihood estimate. i.s also considered. Estimation of variance of the two estimators In addition, ,~e check on things like c~e rate of convergence of the disCtibutions of the two basic estimatoR to normality, and the precision of approximate confidence intervals based on them. ,~c First correct existing series formulae for the lower order moments of the two estimate:l and estimates of variance and give sorne extensions. Gee how things work out in some examples. Simulations ar~ Then we used (when necessary) to check on the adequacy of the series approxi.mAtions. ~n two of the three examples that we discuss, jackknifing can be considered to be an improvem ~l\t on the n,aximum 1i.kelihood estimate on the whole, though the degree of improvement varies a lot depending on the measure concerned. In.the third illustration jackknifing al~aYG improves bias and usually V8t'ianCp.. t.h~ But there is little \lias in the maximum likelihood estimatr.l anyway and p" ..eentage changes in variance are usually small. Tukcy's estimate of variance turns out to be i.nf(~dor In all thre.e examples to simple alternatives. - 1 - 1. Introduc~i~ Despite the recent increase in the number of publications on the jackknife relatively little ~\lantit~~_~~~ even now seems to be known about its performance in less than large samples. This paper should help to remedy the situation a little since its main purpose is to illustrate in numerical terms what can happen to the lower moments of a maximum likelihood estimate for a single parameter after it has been jackknifed. We also spend some ti.me considering estimates of variance of maximum likelihood estimators and their jackknifed forms. and the problem of setting confidence intervals for the single parameter. The only prior publication on the jackknife that seems to be directly relevant to our study is that of Brillinger (1964). We have elected to nlake comparisons of maximum likelihood estimates with their jackknifed fOl;ms via the t:'ornents of their sampling distributi.ons. but recognize the limitations that this approach imposes. The exact moments of these distributions are nearly always impossible to derive. this is so we use two lines of attaC'.k; When firstly. we approximate them by the first few terms of a series expansion and secondly. we calculate them from computer simulations of the problem. Using these simultaneously (or the exact moments instead of simulation if they are available) also gives an accurate picture of the snmple size needed to validate the series expansion in practice. Let us be more specific about our objectives. The jackKnife is a procedure that hes been shmm to have certain desirable properties for a Hide range of problems and been conjectured to have many more. resolve some of these conjectures we will be addressin~ In order to help ourselves here to questions about jackknifed maximum likelihood estimators like the follmving: (i) (ii) how much bias does the jackknife eliminate. i! any? does jackknifi.ng reduce variance and if not. is its net effect on mean square error favourable or not? - 2 (iii) does the jackknife cut down the third moment of the sampling distribution, and what effect does it have on the fourth? (iv) how does Tukey's well-known estimate of the variance of the jackknifed estimate compare with the standard one and how satisfactory is it as an estimate of the variClnce of the .. maximum likelihood estimate itself? (v) does the distribution of the standardised jackknife estimate tend more quickly to normality than that of the maximum likelihood estimate? (vi) which estimate, maximum likelihood or its jackknifed form, leads to the better confidence interval, and how do they compare with Tukey's method for setting confidence limits? since the jackknife is a procedure with high l;>pplied potential it scems to us that questi.ons like these ought to be resolved. Simulation of the problems in this paper have presented a few challenges that h:we been overcome. Second-order moment formulae in principll~ should have been a minor problem, since almost all the ones we need aI'e summadsed in Bd llinger' s paper. Unfortunately, however, there do seem to be a few minor typing errors in that paper and we have thought i. t worthwhille recording here what we believe to be accurate versions ot these fo~~ulae. In particular, these are needed if some of our recommendations are to be adopted in prClctice. 2. Notation We shall suppose that we have a random sample of si~e n, Xl' X , .•• X , 2 n from a single-parameter distribution with density function f(x; 0). This function will be assumed to satisfy certain regularity conditions that will become apparent as they are used. To add to the Ben~rality the basic random variable is taken to be vector-valued, but since the form of the algebraic n - 3 - results is the same whether X has one component or ten, we will not hother to use the conventional bold-type lettering. In order to condense the algebra of the next section we will denote f(x.; 1 e) by f. 1 [a g f i ) -E -DZ--· by I E (a3l0~. f i ) by J E[~~~g fi ] by K E[~lO~!.] by L 2 10 ae ae 3 ae 4 aet. 1~~S-) ae n by a. and 1 l r,og f, ~_. _ _ 1 l: a i"l ) + I . hy b. and 1 af)2 ) {:'l~,:; Jt by c. arid K} by d. and r':.".·s t} by e. and ae 3 ) {'"og f; ae 4 ae S by A i n b. by B }; i=l 1 n 1 c. by C ~: i=l 1 n 1 d. by D L i=l 1 n 1 (1) e. by E l: i=l 1 Evidently all of the b's, c's, d's and e's have zero means; f(x; e) is such that E(a.) .. 0 also. 1 l~e assume that In addition, \.e shall denote (1,11)· (.a_:~_:_~!.._~) } by (1, 1, 11, 11, Ill) (2) and so on. This is simply the single-parameter version of a multi-parameter notation used in Fryer & Robertson (1972). - 4 - In this paper, the maximum likelihood estimator of e, a , is taken n a log f. to satisfy th e equa t '" .. 0' As far as the J'ackknife is 10n • t.. ao 1 A la-a 1=1 concerned, we shall suppose that the basic data is divided (at random if r < n) into r separate, equal-sized groups with n '" rs. a when estimate of The the i th group is omitted is denoted e(i)' notation the jackknifed estimate, 0p' is then defined by ~~ximum Using this eAp .. {A rO - 0 IS 1 _ r r A 1: e • Where the epi = {rO - (r-l)e(i)} A likelihood 1 r A } (1--) 1: a . r . 1 (1) 1" A are the usual pseudo-values. i=l p1 Again, in order to simplify the algebra later on we use Aj to denote1:a i where the summation is taken over membeu of the jth group, with a similar meaning for 3. and C., D, J J B" J E •• J Homent Expansions. lve start 'i':'th the maximum likelihood estimate, a. The local Taylor li a log f.1 expansion of 1: in a neighbourhood of e gives i=l Ie'" a n a log f. (B - n1) + I (8 - 0)2 (C .,. nJ) .. 0 .. A + (0 - 6) . r ------~ ao --ae-- A A A A lO=6 1=1 .,. ~ (6 - 6)3 (D .,. nK) .,. ~4 (6 - e)1I (E .,. nL) .,. ••• (3) Inverting (3) about the origin, we find that (0 - 0) .. where Cl .. L .,. 0 [An55] !:...,. ! .,. L + n °n2 n3 nil A I II .. AB .,. A2J 12 21 3 (4) .. - 5 - (5) . 2r r 2r+l In general E(A ) is O(n) and is of the same order as E(A ) and similarly for other terms. To find series expansions for the moments of 6 about 6 we now have to evaluate terms like E(A 2B). All of those that are needed for the . order we are considering here arE! listed in Appendix 1. Using them and the symbol Ese to denote the series expansion of an expectation (and Vse for a variance) we find that Ese (6 - 6) .. niz + + . 1_ [(11, 1.) + pJ 1_ [(11, II, 1).+ !(111, 1, 1) + n 21 3 n 2 I t<l- [1. J(ll, 1, 1) 2 l 1 Ese n J2(I, n21~.' _L (8 - 6)2 ... -! n1 1 I, + n2111 [J (I, E se (8 - 6)3 A E se (6 - 6)1j 1 + --!n 213 1) + 6(11, ["_2 (II, I, 1) + 3(11, 11) + 3(111, 1) + [(1, I, 1) + 9(11, 1) + .. + 0(n- 3) 3 KJ 1)2 + 12J (II, 1) + 154 J2]. + 0(n- 3) .. Iifj3 ~ i J~I I, 1) + 6J (11, 1)2 + 15 J2(ll, 1) + 15 J3.1 + 0 (n- 3) 2 8 -~ n~1 i ~1 +-6 R (1, I, 1) + 3(11, 11)(11, 1) + 3(111, 1)(11, 1) + 3J(11,11) + 2K(11, 1) + 3J(111, 1) + + (Ill, 11) + !(1111, 1) + i J] + O(n-3) (6) These results, which cover both the discrete and continuous cases, check with those of Bowman and Shenton (unpublished Oak Ridge National - 6 - Laboratory Report) which were derived for the discrete case only by a very indirect method. Further terms in the expansions are not a practical proposition at the present time. Turning now to the jackknifed estimate. 0p' we see that an expansion A for (0 p - 0) is easily asse~bled from (4). For the present time we will regard the number of groups. r, as being fixed as n increases, and consider A other possibilities later on. The following expansion of (9 - B) p is A enough to determine the seeond, third and fourth moments of 0 about 9 to . -2 • P order n A (9 P «* + r8* - 9) • --- + n 2 where + ••••• (r-l)n A «* . 1 8* • t2 (AB - +..L [A2 - 213 yot. ~AjBjJ J !3 EA.2] j J. r(l + !)r AB2 - I' + EA.B~ + 1(1 + j J J !) r AEB.2 - 2BEA.B. jJ JJJ A2C - AtA.C. j J J ICEA.2 + I EA.2C.] j J j J J 1 [,3· 1 2 + -rtt -(1 + -) JA B - 3JA EA.B. I 2 r j J J - .. 1 (1 + -) 1 -3 JB EA. 2 + -3 J EA. 2 B. + _. 2 . J 2 . J J 6 r J I J KA EA. 2 + ! K EA. 3] j J 6 j J (7) - 7In all of these terms the range of summation for j is from 1 to r. The bias in 8 p can be found by linear operations on the bias for 8, but to save space we omit the details, and simply quote the final result. To derive the moment expansions for 8 p we now need to evaluate terms like E[A 2r.A.B:] , the maximum likelihood counterparts of which j J J" were given in Appendix 1. All of those needed for the expansion (7) are set out in Appendix 2. Summarising, we find that r MZ + 0(n- 3) where M2 Z (r-l)n 2 of the n- term in E (0 - 0). se E (8 - 8) = se p r [(1, E se (8 p + is the coefficient r (11, 11) 1, 1) + 6(11, 1) + 3J] (8) - 8)4 Because there are usually fc\·;er terms in the moment expansions for than there ure in thei r maximum like lihood counterpart~. ep it is tempting to conclude that the moments of 8p are therefore smaller in absolute terms, but this remains to be seen. In these expansions we have assumed that the number of groups, r, is fixed. However, some might argue that it is more natural to fix s, the nunmer of observations in each group, and let r ~ ~ as n ~~. For this reason we have also traced the algebraic steps holding s constant. - 8 - It turns out that the results for fixed s to order n-2 can be very simply obtained from those for fixed r. All we need to do is to set r = n in (8) and retain terms in n- l and n- 2 simply omit the factor (r 2rr Equivalently we can in the expansions at (8). Only if we went on to consider terms of order n- 3 would fixed s show itself clearly. Of course we are not restricted to fixing r or s as n increases either we could take them both to be proportional to n l for example. However, formulae in this case are·much more difficult to derive and we will not attempt to deal with this case here. It j.s probably worth mentioning here that several checks that we have applied to the jackknife results at (8) have turned out to support them. For instance in the case of a Poisson distribution with mean 6 when both 6 and are the sample mean, the appropriate terms in the 6 p moment expansions of a p tha~ sum to zero as they ought, despite the fact terms like (II, 11) and (Ill, 1) are non-zero. We also evaluated the moment expansions in a couple of very simple cases where we knew that the second order terms did not sum to zero. Estimating a in the negativE:; -ax exponential density, f(x) = ae , x > 0, gave for example E se (a a2 p 2r a2 - a)2 =-+ n n 2 (r-l) + ••••• (9) compared with the moment of the maximum likelihood estimate E se ~ 62 (6 - a)2 = -n + (10) Again, estimating a when the data come from an N(a 2 , 1) distribution leads to E se (ap - 6) = 1__ + r 2 4na 32(r-l)a 6n 2 + ••••• (11) compared with E se (6 _ a)2 = 1_ 4n6 2 + .......;1;.;;5 _ + (12) - 9 - Since the expansion for the variance of 0 in the first of these examples is (~~ + ::~) and (_1_ 4n0 2 + _-lL + ... 64n 2 0 6 J in the second, these are both cases where in the series sense jackknifing reduces both the bias and the variance of the estimate (r '" 2 in the first case actually gives equal v.;triance). This also seems to be the appropriate place to record moment op • The standard expansions for estimates of the variance of 0 and procedure for e~l:imati.1g the asympt;otic variance of ~I and substitute 0 for 6 • is to take the form v.'e denot!> this stad stic by VI and this also ~ serves as an estimate of 0 " vee p ). Substituting Vee) and vee p ) which . 1 0p for e 1n:oy gives an ~ alternative estimate for we denote by concentrate her.e on these two estimates together with r estimate of v.dance, S2 = 1) I: (0 • T r(r-1 i"l pi ~ ~ 8(i) - special circumstances. (r-~ we1J-known There ar.e other estimates P r(r-T)" re 2), but thesE:'. only' seem likely to work \,'",11 in i=l r Tuk~y's 1 which might be used, for examp'le, and J!-l)_ e )2. V2 • p D.R. Brillinger (1966) has suggested the use of r (L 02(i) ra 2 ) in certain situations where the bias of a is small. i=l As far as maximum likelihuod estimates are concerned ,,,e can. ShiN that r E r (r-l) se l.--r1 where - nI2 r (r. i=l e~i) J".I [0.1, 1) +2 ~ 1 w_ [(11, 1) + nX I] + O(n -2 ) (13) is the first order term in the bias of 0, and this of course squares with his suggestion. Using E se (VI) expansions for =·L+L {L r!+ nI n 1 [2 2 3 e and (111 0p ' given previously we find for VI that 1) + (11, 11) -T--- + (11,1,1).1 - 2 (4) - 10 - E A se (V 1 - 1 )2 1)2 + 2J(11. 1) +(11, 1)2J nI As far as V2 1 is concerned Ese (V2 - __)2 nl A Ese (Vi) =L+L nI n2 { 1. 13 .. E (V se 1 + O(n- 4 ) (15) to order n-3 , but - 1__ )2 nI 11) + (11. 1, [!2 + (111. 1) + (11, 1)J --r--2 (16 ) The moments of 5 2 are a different story, however. T A straightforward but very tedious piece of algebra for fixed r produces the result + __.I..-n 2 (r-1)I 4 [21(11. 1. 1) + (1, 1, 1) J + 21(11, 11) + 4(11, 1)2 + 31(111, 1) + 9(11, 1) J + IK + 3 J~+ 0(n- 3) (17) . r(2r-3) . r and replaclng ----- by 2 and - - - by 1 in this expression gives the (r-1)2 (r-1) (5 2 E T se - + 0(n- 3), giving it zero estimating efficiency L )2 = 2 nI n 2 1 2 (r-1) A in the limit compared with V and V • Using fixed s on the other hand 2 1 ] -3 reduces E (52 - ~_)2 to order. n which presumably adds to the case for se T nI using fixed size blocks. ~stimating Incidentally, several simple checks (for 1p1e eX811 the mean of a Poisson distribution) applied to the formulae at (14) to (17) again proved to be positive. Finally in this section a word about estimating the variance of e unbiasedly to second order. e for In view of (14) we cannot simply substitute e in E { se (0 - e)2 - r-<11. 1) + JOJ2) 2" f where E (e -e)2 is given at (6). se ~ 11 - . We have also to subtract the second order term in Ese(V l ) and when this is done an unbiased estimate becomes { ~l'-' ' 21 1 n 21 3 + where of [~+ 2(111. + _1_ .n 1) + ~ (11. 11) [J(l. 1. 1) + 3(11. 1)2 + 1_ n 21 4 ... e is substituted for . .. e. +l J~ (11. 1. 1)'] J(ll, 1) + 2J2]} (18) Similar remarks apply to the estimation Vee) using ep and tC" Vee p ) using either ep or o. 1-Je could also consider es timating e i taelf along the same lines. sinc.e on substituting in {8 + n~2 I: (11. 4. ~ome 1) + ~ ]} we get an unbiased estimator of 6 for e e to order n-1. Illustrations We now turr tc Borne concret.e estimation pr...;'lems (which· were selected ~ore for ease of computation than for their intrinsic interest) to show how things can work out ...umerically in practice. Before 100Hng at the details the-I1gh, we ought to make it clear from the outset that ,,,hilt 'Ne are doing is con~aring the certain situations. It perfo~lence ~oes of the t~o estimators, 8 and . 8 , in p not follow that we would recommend the use of illper of thflm in any particular instance, since 8. completely di fferent kind of estimate may be clearly preferable. ~xamp~e 1. We suppose that X has the special beta density f(x) • (8 + 1) (6 • 2) xe(l - x) and that we want to estimate formulae for 6 and alog f ae I • -a 2 { and 8 > - 1 p we need the following facts 1 • < 1 + (2+6) 1 + log x I} (1+e)2 + (2+8)2 < 1 In order to calculate the moment 6 • (1+8) 10,8 f, ae 2 6. x where 0 (19) -12- L = 24{-_1__ (1+0) -S + __(2+0) l_-S} = 0, (II, i) = (Ill, 1) From this we can deduce that (II, I, 1) = -1 2 and so on. ShOl~ that t It is qui te simple to a single sufficient statistic for e n 1: = 1 n i=l log x. = log. G say is 1 and that the basic density is a member of the well-known Koopman exponential class. The maximum likelihood estimator for 6 is given by (20) and this is also sufficient. Clearly jackknifing here will destroy the sufficiency property, but this may not be such an absurd thing to do if we cannot find a oneto-one function 0f log G ",ith suitable bias ana variance. have said before ,~e In any event as He are using this estimation problem only for illustrative purposes. First let us look at: the results for bias which are given in Table 1 for selected va:ues of nand O. ~~ote that in the series vC\lues the contrihution of . -1 tlw term 1n n completely dominates the bias of jackknifed estimate is virtua)ly unbiased. 0, and that as a consequenc<' the He have thought it sensible to si.mulate the problem too in order to check on the accuracy of the series results. e given simulated results for The in Table I appear to be pretty much in line ",ith the 'i'eri.es results, though a very large number of generated samples is needed to establish this. For obvious reasons even Dlore simulation runs are needed to check on the accuracy of the bias of iable than those for 0 beca~se e. p Although the results we have here are less rel- of the prohibitive number of runs needed, they do confirm the order of magni tude indic.ated b)' the series formula. In all cases the difference between the series and simulated values is covered by two standard errors of the simulated estimate. All of the results in Table 1 assume that r is fixed and set equal to the prevailing sample size and this is the optimal choice for r. Usin; the series formulae for fixed s ('" 1) gives very similar results but they tend to be slir,htly further al·ray from the simulated values than those for fixed r = n. -13The corresponding series Dnd simulated calculations fer the variances and mean squared errors of the estimates ore also set out in Table 1. Note that in the sed es results the firs t: order term is dominant unless the sample sbe is very small and that jackknifinp, which always reduces variance, leads to error in the smaller sized samples. ust.~ful gains in meen squBl"ed All of the simulated values sug~est that the series results in Table I are close to the true values of the second order moments. We have BRain quoted results for fixed r = n. Calculations of both series and simulated second moments for other values of l' indicate that and this is what ",e We should ah'ays use as many groups as possible, would expect to happen in general. The series results for the third moment that we have derived only to order n -2 E give the ratio (0.'0) 3 se ... _._-_ E se ~ - a value of 7 1 , The simuletionc. 4 (0 -0)3 p however, give somclvhat different ratios ,,,hen n is small and really qui te diff~rent values for the i.ndividual third moments from the sedes results when n < 20, es ",e can see frem 'fable 2. entirely unexpected. This of course is not a \-lhether the measur.es of skewness for Rnd similar or not depends on whether the moments are taken about respective means. Taken about E(&) or E(e ) the p especially when n > 6, but the picture if; very the location, because of the bias in e. m~asures diffen·~t a p an" e or about their are very similar if (3 is used as The results for the fourth moment displayed in Table 3 show the same trend but are less pronounc:ed. Judging by the calculations in these tables for the third and fourth cumulants the distribution of 0 p normal density about distribution of O. a can be adequately approximated by a for much lowe l' values of n than can the Again in both tables fixed l' sample size since this produces the best results. s = I give very similar values. is set equal to the Results for fixed -14- MovinE on now to estimates of the variance of 6 and 6 p' we find some numerical values for expected values set out in Table 4. -2 The simulated results indicate that the series values to order n adequately represent the moment in question. and set equal to n. Note that E(V2) < E(VI) for series and for simulated calculations. In all cases r is fixed E(Si) invariably both < Co~paring the results in Table 4 with those in Table 1 we see that VI gives the best estimate of ~ V(a) for both the geries and simulation results. si appears to have least ~ bias for estimating E(6 - 6)2 using the series results, but comparable bias with Vl when the simulated results are contrasted. standard error of Si is roughly double that of VI and this ratio goes Vl has slightly less bias than V2 up rapidly as r is decreased. (which is always closest to ni ) as an est~~ar-e of E(a slightly larger standard error. here for using V1 However, the p - 6)2 but There is clearly a case to be made whichever of the three measures Wi' wish to estimate. Finally for this illustration we :;ive some numerical comparisons for confidence intervals. We are mainly concerned here with comparing the probabilities of covering the true value of methods are used. Method I uses the nominal 100(1-a)% level interval , whe!"e normal density. e when three approximate Hethod a is the upper 100 2" % point of the standardised similar and based on ep -+ ~CY1 72 Hethod IlIon the other hand uses 6 + p- t aT> (r-l) number of groups used in the calculations and t a ~ .rv: where r is the is the upper 100 2"' (r-l) point of the central t distribution based on (r-l) degrees of freedom. There are other variations that we might have tried like r;:-' ~ or [a .:!:. ~a Y v2] [eP-% + ~ ~l] - (or even modifying the number of st<l.ndard errors used) 2" but the line had to 2a % -15,- be drm·;n somet~here. Estimated c("Jnfidc\lce c.oefficients (,dCh standard errors of generally less than set aut in Table 5. 1J %) for various values of 0 and n are Method I seems to intervals (given only for 11 .. eive sli~htly too many 'low' 20) but is othend.se satisfactory. Hethod II i.s poor for smaU samples but improve.s cw 11 increases. For n .. 10 the confidence coerficient is some 3-4% too low and there are far too many 'low' intervals. In view of the fact that ep is virtual!y unbiased and the comparati ve rest~1ts for Sketmefls and kurtosis this is something of a surprise. It is true that V2 is biased downwards for V(fJ ) and so replacing it by V1 \"hich is biased tlp\mrds mllY p improve the ovetallcoefficie.nt. were l"(~gistered nut even if all of the improvement in the '1m·.. ' interval area, it 'vouid still produce a use fixed r .. n bl~ttcr ,,'ell balance. seems likely thllt Nethod r All of the results for Hethod II other values of r mere'y make thingr- WOl:'SC. Method III seems to give satisfactory overall coverage but the split between too 'hieh' and too 'low' intervals is won:;'i!:g. again gives the best results. coeffi.cient, il1t~r.va1.s Fixed r ... n In term:, of physi.cal length fOl: a nOI:111::.11 generated by Hcthod I ar.:- some"'hat longer than those for Method II but consi derably shorter than those pro(!llced by Hethou III even ,~hen n .. 50. But this is only to be c,:pectecl in vie'" of our previous results. pI q Our second illustration concerns the estimatioll cf 'i' ~ (rather than the more usual lof, odds) t"here p i.s the probabU:: ty success in a llernoulli. trial and q ... 1. - p. In terms of <1. of the frequency function of the number of successes, X, in a single trial is simply !pX f(x; cjJ) .. Ti+~T for x .. 0, I . (21 ) -16- The maximum likelihood estimate of ¢ in n trials is given by A cj>.= o/r for k = 0, I, 2 ••• (n-1) TF-Tr) where k is the total number n When k = n there is no maximum likelihood estimate, but of successes. if we define cj> = c kin here where c > (n-l), then viewed as a whole ¢ is a one-to-one function of kin' which is a single sufficient statistic for 9, and is therefore a sufficient statistic for ¢ itself. obvious problem with value for c. in cj> pr~ctise is that we have to nominate a suitable To avoid generating several sets of results for different values of c we have decided here to make comparisons between jackknifed estimate, distribution in the cj> p , upp~r appr~aches and the cj>, using a truncation of the basic binomial tail. ~~en the truncation is light, is not too large, n is not minute and c is not outrageous the from the two The cj> co~clusions will be similar. Of course we have the same kind of definitional problems with ¢ • p As we reduce the valu~ of r so this difficulty becomes more acute and the further we have to truncate the underlying distribution to avoid it. But the higher the degree of truncation the wider the gap becomes between the properties of the Lppropriately defined estimate for the full A distribution and those of ¢ over the truncated distribution. p At the same time ¢ becomes less and less a maximum likelihood I"stimate (the subject of this study). For these reasons we use r .. n only for simply omit the points k = n, (n:-1) from the sample space. ep and This turns out to be one of those cases where jackknifing with r .. n does not greatly complicate the form of the estimate and it is very simple in fact to show that ; = ; [ 1 - (n-ll.p n(n-k-l) J for k .. 0, I, 2 ••• (n-2). A Note that 0 < ¢p -< ¢ . By suitably defining ¢ for k .. nand (n-l) p A (and changing it at k = (n-2» we find that ¢ is sufficient for P cj> too, -17which is not what usually happens when n sufficient statistic is jackknifed. ~ The value of p at k = (n-2) needs changing for two A reasons. Firstly, even though ¢p is not continuous it can destroy the monotonic nature of the estimate and with it the sufficiency. Secondly, ·and more important, it can lead to an outrageous estimate of ¢ with high probability. In this case ¢ ¢p 97 = ~. = 49 Consider for example n and ¢p = 0.49 though when k = 100 = 97, ¢ and k = 97/ 3 = 98. and l{hen ¢ is very large we will tend to get large values of k so in this situation;with k =(n-2),¢p will produce a disastrous estimate. A In practise we might replace it by ¢ or ¢ that we comment on later. *= Of course using; k (0":"j{"":i1Y , an estimate p =~.n at k =: (0-2) when ¢ is small will alrr.ost certainly be a point in its favour but the probability of observing that number of successes will then he negligible. Beca~se the value taken by ¢ p little effect on its moments unless ~ will have very is large, we have not adjusted it in any way here since we always take ¢ takes at (,,-2) wi 11 not affect. the = (n-2) at k ~ 2. mom~nts Of co~rse the value that ¢ \olhen ¢ is large ei tht)r provided that we can then p~rsuade ourselves that it is ~ that we are interested p T~ere is no need for simulation of the moments of in rather than r!! q these estimates in this binomial example since the exact values are easy to compute. computed Series approximations are ..mncc(,F.sary too but we have them (on the asslwntion of a full binomial distribution) for several values of nand merely to see if .md ,,,"hen they adequately cP represent the exact values. Before looking at the details of the sampling moments we briefly . .. . d1scuss a thlrd estlmate cP * n = (n-k+l)·' This estimate results from questionning how one should choose 0 and £ (k+£) bias of -(n-k+6T the bias of this estimate can Provided that 0 > a p to minimise the first order -18- [(l:~) be written as the true value of <I> which eliminates with ljl P { ~(l-~) + ~} + 0(n- 2 )J. Since we will not know ~ in practise the solution must be first order bias entirely. =1 But how does and 'I< <I> E =0 compare ? The typical pattern of behaviour of the bias in our three estimates is shown in figure 1. We have used ljl more distinct as we raise the value of bias for ljl = 0.5 of the bias in ljl =2 there because the graphs become ljl. Additional information on and 1.0 is given in Table 6. The initial behaviour is peculiar and quite unexpected. First the estimate is biased downwards. then it passes through zero to become biased upwards. It then reaches a peak when n is still small and finally tai 18 off to zero. The series approximations on the other hand are positive monotonic decreasing functions of n and they usually beeome tolerably close to the exact v8lues after the initial peak in the bias has been reached. The behaviour of the bias in both is quite different from that in cp. ljl p and ¢ 'I< TI."y are always biased dO\vflw3rd and their bias tends monotonically to zero as m' increase the sample size. It came as something of a surprise to us to f:nd that the bias in ljl 'I< is much smaller than that in ¢. p improvement on ljl. b~th However Putting the initial behaviour of ~ are a great to one side it is clear that as we raise the value of $ so the bias in the estimates takes longer (in n) to disappear. just as we might expect. The patterns of variance of the three estimates are similar however, though the absolute levels are Quite diffl!rent as ""e see from figure 2 and Table 7. First there is a rise in the region of small n, then the variance decreases to zero after showing a single peak. Until the first order term becomes dominant (so to speak) which takes longer as $ increases, " -19- the variance of ¢ than that of q,. is much swaller than that of ¢ * :md vpry much smaller p The key question to ask now is hOV7 thinrs '\omrk out the bias and vari ance are put tor,ether to fonn lTlfWn ~~hen squared error. There is an initial region for very small values of n where q, has the least mean squared error (because of the behaviour of its bi.as). of its bias ber,ins to '''ane the lead changes hand and $ But before long the smaller variance of cf> p p and cf> * after * is pref"rabJe. outweighs the smaller bins in ¢ and the gap between the two becomes considerabJe. way behind both ¢ 1'.0 the dominance Evidently cf> .;, lags alonr.; its early dominance, so urtless n is very small or very large there is little to be said in its favour. distrihution of ¢ p converges to normality about ~) mur:h L,ster than does the distribution of $, if the third and four eh t::-mer,f:r. <::re to be the criteria. This can be clearly s£'en from Table 8 "There valncb of these highcl' moments are shmvn before and after standardisl!don. 'fhe standard estimates of the seconl order moments of l' and we have previously dEmoted by VI and V2 (')2l0g f) ·---;;2- take E ( 'p that have ver:" simple forms j f ...l e = E {(1+¢)-2- k<P- 2} over the full binomial distribution, Lut considerably more c.omplicated functi.ons if we take into account. II> t~le Since we takE' <b to be relatively small the betHeetl the two for any particular est:irr,·~te truncation llun~0ric:ll will also be sma]], ,ud have opted to scrutinise the behaviour of the simpler estimates. diffe-rence ~o I.e The moments of the Viand s~ though are tak{,n f)VeT the truncated distribution usetl previously since otherwise ','e have the same defi.nitional problems as before. Very little effort shows that VI substi tute .p for $. ~ _;(1+;)2 ....._,-.. and n Sli.ghtly more algebra produces s~ -20- Clearly V2< Vl (when k t 0) anrl Vl < Si unless k is very small. Note S2 T .~.-' that a 4 when k = (n-2). Because of these relationshi pro the Vl results in Table 9 ~o1ill come as no surprise. tends to over-estimate Vl both V(q,) and V(4)p) Hhilst V2 t£mds to over-estimate V(cf>p) but underestimate V(q,). si, on the other hand is badly biased upward for all of the second order moments unless n is large (due to its b£!haviour for large k presumably). Furthermore the standard error of Sf is often considerably larger than that of Vl and frequently several times the In this binomial problem with the size of the standard error of V2 choice of these three estimates we would use V2 to estimate V(¢ ), P Vl to estimate V(¢) and discard Sf altogether. Finally, for this example, we have some coverage probabilities and expected lengths of the approximate confidenl"c intervals referred to previously as methods I, II and III. We could develop exact confidence intervals here of course but this is not the point of The length of each interval is zero when :, = 0 but ~·]e ~he exercise. feel that this is inappropriate for an interval based on the normal distribution (or t) and so ~""e have also conditioned out this point. All three methods tlsull11y g:i,ve acceptable overall coverare levels "'hen n > 20, as we can sel' from Table la, but the split between the number of too 'high' and too '10,,,' intervals is very bad for each interval even for n = SO. The accuracy of coverage for method II Seems to worsen more rapidly than for the other t~o1O intervalG as we raise the value of ~. This is a pi ty because its expected length for the same coverage probability is usually shorter than that of method I and often considerably shorter than that of method III (c.f. Table 9, also). This concerns the estimation of the parameter A in a pO\o1er series distribution'which has frequency function -A x f (x) = _L...L_. (l-e -A)X! for x = 1, 2. 3, ... (22) -2]where ~ > O. ~, ~, The maximum likelihood estimate of is found from the equation -x (1 - e and since -~ x is - ) ~ =0 (23) sufficient for ~ the same is true of ~. There is an extensive literature on this type of estimation problem and a considerable amount of work has been done on this special case of estimating ~ when the zero count in a Poisson distribution is ruled out. The minimum variance unbiased estimate of ' A, ~ (which is a multiple of i), has been explored by Tate and Goen (1958). The multiplier for xbasically involves Stirlings numbers which means that it is not simple; however Tate and Goen provide tables for it and a recurrence relationship for it has recently been given by Ahuja and Enneking (1972). Ad hoc estimates of ~ have been given by David and Johnson (1952) and Plackett (1953) among others. best of these is Plackett's unbiased estimator number of elements in exact variance [ >. + th~ ~ * . .. rn 1'=2 n sample taking the value r. >. 2 (e >. - 1) -lJ In, The r. --~ where n is the l' This estimate has and has been shown to be about 97"1. efficient against the Cramer-Rao bound when>. • 0.5, reducing to a minimum of 95% or so at ~ = 1.355 and then increasing to 100% as >. + Tate and GO. * 1 ~ 1 Goen also show in their paper that nl < Vo..) < V(>. ) where nl is the usual Cramer-Rao bound. for using for ~ * if In practise there is evidently a strong case zero bias is required, even though it is not sufficient L However it is possible that the maximum likelihood estimate ~ will nI . 1 . h ave variance or mean squared error below and that Jackknifing will ~ this is one of our reasons for looking at the problem. We have computed and ~ *, reduce it still further to give worthwhile savings over and the usual series approximations to the lower sampling moments here and ·-22- and have also checked their accuracy via computer simulations. All of the results for the jackknife use fixed r ~ n and this is almost invariably optimal. The elements needed to compute the se....ics approxi.mations are easily Mrived, but 11luny are quite lengthy in appearance. For t'xmnple . (1 + 3>' + (1, 1, 1) .. ··..·-····--::.r··A2 (1 - e ) + •. .1 . (3e- A(1+A) - (3+A)] + ---_•.-.• '.-•.'----'------'-"'- A2 ) A(l - a-A):! ._ 4(1 + A) -------:.::~-- A2(1-e ->- )3 A3 (1 - e )2 and so to save space will not eive the moment formulnc in detail. Before turninr. to the numerical resul ts, hmJever. there are one or two minor points thi:t we ought to make about these estimates. = 1 for all i then A .. 0 ncte that if Y.. of calcul..tt~_tlg ;I.. Hhen 1 the point to avoid complic:at:i.ons. vBI~es as well a~ l1cnts of ), l!l(lI Firstly we which is an 'illegal' estimate '~e have included this sample It is also possible for \ t~ take negllti ve zero (thourh the probability of this will usually be very small inceed) and thi s has b(>en treated in a sit:1iJ ar '''ay. note that Sf =0 Finally "ye if the means (If all the sub-groups (",hen jackknifing) are the same, and confidence intervals of zero length have been excluded in the simulations. Some series bias results for n =5 ~re set out in Table 11. Evidently the first order term for A is dominant leaving the jackknife virtually unbiased, but the overall level of bias in A is small (less than 5% of value of A) -23- even when n '" 5. So although jackknifing reduces the bills by 95% or so, the total saving is scarcely \~orth th~ effort, and as n incre,1ses the Being so very small the bias in A is bias rapidly approaches zero. p difficult to establish by simulation, but the results that \-1e have tend_ . to support the series level for it. Simi.lar remarks app 1)' to A• A Some calculations for the second order moments of A and A when p n .. 5 are given in Table 12 and they are typical. The results are disappoi.nting, principally because of the dominance of the first onll'r term. The mean lIeJuared error of A is always above the Cremer-Rao p bound and exec.eds both the variance and mean squared error or A for small values of ),. By tLe time on both the bias and variance of E fA -A)2 E (A-A)2 From then on ---~!':....'-!-_.. se A ~ 2, however, A is cutting down p A, thoueh by very li ttle in 0.99 as A increases. the behaviour of A is worth commenting s~cond order terms in V (A) and se r. se ().-).) overall level::: below the Cramer bound. over A or A* is some 13% at A c 011 ,~hen :lb50 1 ute terms. It does seem to us that A is very small. Here the are negative, so takinr. tll" The 0.1 when n efficien~y ~ improvement in A 5 and may be worth the effort (perhaps it is Horth seen-ching for the minimum mean squared er.ror estimate in general!:). Sim'Jlatcd results that we hnve [or thef:e second order moments suggest that th~ series values may be a frac::ion 10'-1. In view of the size of the second order moments of A and A it may l' be argued that there is no longer Bny point in continuing our study of this problem. I!owever, as we said at the beginning our main objective is to compare A and Ap ' not to contrast will comment briefly on some t~ern with other ~stimates, so we other features of our two estimates. Series calculations indi cate that in the smaller siz(!<! samples the thi rd moment -24- of Ap about A is almost always at least double the size of that of ). and often much more. Simulations for this third order moment which we believe to be quite accurate show that E(). p - ).)3 > E(~ - A)3 also, but indicate that the gap is nowhere near as wide as the series calculations suggest (usually less than double). Standardising this moment by dividing by the appropriate power of the mean squared error leaves these conclusions virtually unchanged. It may be worth noting that these third order moments are very much more in line when taken about the respective means of the estimates. to order n ··2 Series for the fourth order moments are identical of course. The simulations that we have run here suggest that the leuding term in the series is a very good approximation to the true value of the momen~. Standardising them (using 3 ] typically) shows that judging by the size of the fourth cumulant: the distri1>ution of A converges much faster to normality than p does that of A and likewise if the moments are taken about the means of the estimates, but we must remember that for third order moments we came to the opposite conclusion. We have also run out a few ;;i.mulations on the distributions of VI, V2 and s~. The tentative conclusion from these seems to be that E(Vl) ~ < E(V 2 ) < E(S~) and that all three are biased dm''T\wards ~enerally though not usually by much. Also, although s~ has the least bias, as usual it h~s the largest standard error, it frequently being ten times larger than that of VI or V2 when n is small. This makes the mean squared error of si several times the size of that of VI or V2 es a rule. Finally a brief look at the approximate confidence intervals. Overall coverage probabilities are generally quite accurate for all three methods, though method I does seem to he slightly inferior to the other two most of the time. For example with A .. 2 and n = 10 coverages for • -25method I are 82.4% (nominal 807,), 90.52 (90%) and 92.00;~ (95%). Method lIon the other hand gives 79.80% (BO%) , 89.847 (90%) and 9/~.10% (95%) and method III 78.16% (80%),89.96% (90%) :mc1 94.46% (95%). The tt)il split is less satisfactory, hOI,ever, and \,hen n = 30 and A '" 2 with nominal confidence level 90%, method I splits the too low and too high intervals as (7.40%, 3.72%), method II is (7.40%, 3.92%) and method III as (6.34%,4.00%). Because E(S,f> > E(VZ} > E(Vl) the expected length of intervals for method III is almost certain to be larger than for the other two for a fix.:'o confidence level. n = 10 and a nominal confidence level of 90% IVl' with A ... 2, foupd the expected lengths for the three methods to be 1.6303 (I), 1.6360 (II) and 1.7871 (III). Raising n to 30 produced 0.9495 (I), 0.9505 (II) end 0.9761 (III). Most of this \York was carried out \olhilst: \'le were visitors at the Biostatistics Department, Cha!)el Hill, :lorth CaroliniI. ";e would like to express our thanks here for support from Institute of Generol Hedical Sciencei; Grant G.M.-12868 \·]hilst: we \-lere there. II' addition one of us (J.G. Fryer) would like to thank the Fels Research Institllte, Yellow Springs, Ohio for SUPP01"t during the S\.lmtT,er I)f 1972. his \·1Ork was carried out whilst lIe Part of was a Vifliting FellOl~ at that Institute. Refel"eneE'S Ahuja, J.e., and Enneking, E.A. (1972) Recun'ence relllti.on for minimum variance unbiased estimator of a parameter of Bri1linger, D.R. (1964). R left-truncated The asymptotic behaviour of Tu!:ey'll general method of setting approximate confidence limits (the jackknife) w'hen applied to maximum like lihood (>s timates. • Brillinger, D.R. (1966) The application of the jackknife to the analysis of sample surveys • .~~~e~!?!y, ~ 74-80. David, F.N., and Johnson, N.L. (1952) Fryer, J.G., and Ro~ertson, C.A. (1972) 'I'he truncated poisson. A comparison of some rnethods for estimating mixed normal distributions. Plackett, R.I,. (1953) ~.i_?.!~.(~~_~i:.l5:.a, 12., 639-48 The truncated poisson distrihution. Tate, R.F., .md Goen, R.L. (1958) Ninimum variance unbiased estimation ~Y-.~E.'!~2:._.!: (i) (ii) (iii) (iv) (v) (vi) (vii) (viii) (ix) (x) (xi) (xii) E(A2) " nl E(AB) " nOl, 1) E(A3) .. nO, 1, 1) E(A 2B) .. n [(11, 1, 1) + 1 2] E(Al\2) .. n [(11, 11, 1) + 21(11, E(A 2 C) .. n [(111, 1, 1) - I») IJ] E(A4) " n(l, 1, I, 1) + 3n(n-l)1 2 E(A3B) "n[(l1, 1, 1, 1) + 1(1, I, 1)] + 3n(n-l)l(l1, 1) E(A 3e) .. n [(111, I, 1, 1) - J(1, 1, + 3n(n"1) 1(111, 1) 1)] J + n(n-l) E(A2B2)" n [(11, 11, 1, 1) + 21(11, 1, 1) + 1 3 E(A3D)" n [(1111, 1, 1, 1) - K(1, 1, 1») [1(11, 11) + 2(11, 1)2- 1 3] + 3n(n'-I) 1(1111, 1) E(AB3)" n [(11, 11, 11, 1) + ,1(11, 11, 1) + 312(il, 1)] + 3n(n-l) [(11, 11)(11, 1) - 1 2 (11, l)J (xiii) E(A2BC)" n [(111, lI, 1, 1) + 1(111, 1, 1) .. J(ll, I, 1) - 12 J] + n(n-1) [2(11, 1)(111, 1) + 1(111, 11) + These formulae r.ce easily elltablished. E (A2 BC) " E {[1: a? i + l: i,j i~ 1 r2..(J Cpnsider, for exa",plE', a. a. ] [1: b. c. + 1 J j 1 1 1: i,j i~ bi C• J .!} • Because the variables a., b. and .::. have zero means \1e can wri te t:Jis as 1:1, E(A 2nC) " E r " E l~ [1: i a? 1 (~-~.~.:-!-i b. c. + 1 1 r 2 g [a 1o as 2 1 l: a? b. c. + 2 1 i,j i+j ~.t + J r] J 1: i,j ifj a. b. a. 1 1 (.a310g_~1 - Jl asS J J C j ] which reduces to n [(111, 11, 1, 1) + 1(111, .1, 1) - JOl, 1, 1) - r 2JJ + n(n-l) I [(Ill, 11) + IJ ] + 2n(n-1) (11, 1) (111, 1) (i) = nl E(I:A.2) j J E(A rA.B.) = n [(11. 1. 1) + 1 2] (ii) j J J E(A.I: A.2) = n(l, 1. 1) (iii) j J E(AB.I:A.B.) = E(tA.R.)2 = n [(11. 11. 1. 1) + 21(11. I, 1) + (8-1)1(11, 11) (iv) j J J j J J + 2(8-1) (11, 1)2 .. (8-2)13] + n(n-8) (11. 1)2 E(A.B.I:A. 2) . E(A2EA.B.) (v) j J It j J J = E(rA?EA.B.) j J j J J .. n [01. 1. 1. 1) + 1(1. 1. 1) + 3(11-1) 1(11. 1)1 • n(0-8) 1(11. 1) E(A.r.A~). E(EA~)2 (vi) j J j J = n [(I, 1. I, 1) + 3(8-1)1 2] + n(n-8)1 2 E(A2.EB~) = j J (vii) n [(Ii. II, I, 1) + 21(11, 1. 1) + (6-1)1(11, 11) + 2(s-1) (11. 1)2 - (8-2)1 3] . + o(n-s) l.'-1(11 • 11) - 1 3] E(A.EA.B~) = n [(11. 11. 1. 1) + 21(11, 1, 1) • J J J + (5-1)1(11, 11) + 2(s-l) (11. 1)2 (viii) (ix) (:;) (xi) . - (s-2)1 31 E(AEA?) = n [(1, 1, I, 1) + 3(s-l) 1 2] j J E(AEA.2B.) j J = T1 J E(A 2 EA.C.) = j J J [(11. 1. 1, 1) + 1(1, 1. 1) + 3(s-l) 1(11, I») E(AC r.A~) j J .. n [011, 1, I, 1) .. J O. 1, 1) + 3(8-1.) 1(111, I)J + n(n-8) 1(111, 1) (xii) E(A r A. 2C.) = n [011. 1. I, 1) - JO. 1. 1) + 3(s-1) lOll. 1)] j J J Again. only simple algebra i8 needed to evaluate thesernyments. For instance E(ABEA.2) = rE(A.3B.) + r(r-l) E(A. 2)E(A. B.) jJ J.1 J J J and terms like E(A.3B.) are available from Appendix 1 with n replaced by 8. J J • ~ e e e Table 1 Some Numerical Comparisons of the Series Representations for the First Two Moments Moment A Value of 6 Sample size (n) A A E (ft -6) to se -1 order n Ese (6 - 6) to -2 order n E (6 - 6) to se p -2 order n Var se & Var " (6) A A (6 p ) se -1 to order n (6 - Var (6) to se _? order n • E 6ho se -2 order n E (6 - 6)2 to se p -2 order n 0.2011 (0.2153) 0.1044 (0.1077) 0.0461(0.0463) 0.1679(0.1822) 0.0915(0.0938) 0.0427(0.0427) 0 2 6 10 20 0.1200 0.0720 0.0360 0.1298(0.1247) -0.0117 (-0.0207) 0.0755(0.0745) -0.0039(-0.0048) 0.0369(0.0377) -0.0009(+0.0001) 0.1333 0.0800 0.0400 0.1867(0.1997) 0.0992(0.1022) 0.0448(0.0449) SO 0.0144 0.0145(0.0146) -0.0001(+0.0002) 0.0160 O.01tiS(O.0166) o.Ol?() (0 .0168) 6 0.3150(0.3248) 0.1833(0.1884) 0.0895(0.0869) 0.0353(0.0386) -0.0286(-o.C267) 20 50 0.2912 0.1747 0.0874 0.0349 -0.0095(-0.0097) -0.0023(-0.0052) -0.0004(+0.0003) 0.9600 0.5760 0.2880 0.1152 1. 2903 (1. 3993) 0.6949(0.7357) 0.3177(0.3261) 0.1200(0.1219) 1. 3751(1.5048) 0.7255 (0.7712) 0.3254 (0.3337) 0.1212(0.1234) 1.1635(1.2515) 0.6438(0.6750) 0.3041 (0.3099) 0.1177(0.1195) 6 10 20 50 0.4582 0.2749 0.1375 0.0550 0.4961(0.4918) 0.2886(0.2737) 0.1409(0.1472) 0.0555(0.0475) -0.0455(-0.0611) -0.0152(-0.0347) -0.0036(+0.0032) -0.0006(-0.0008) 2.4590 1.4754 0.7377 0.2951 3.2891(3.5421) 1. 7742(1.8231) 0.8124(0.8291) 0.3070(0.3073) 3.4990(3.7839) 1.8498(1.8981) 0.8313(0.8508) 0.3101(0.3095) 2.9629(3.1597) 1.6434 (1.6741) 0.7775(0.7899) 0.3013(0.3016) 10 0.0164(0.0163) 4 I Note: The figures in parenthese.s are the simulated values. r is fixed and set equal to n in all cases. As far as 6 p is concerned Table 2 Values of the Third Moment and M ~-~ ~--- -- -J--~.~--"J --_._--~~-,_. -,~"~--,,,., 0.1120 0.2194 0.0640 0.0967 2.1966 1.243.. 10 0.0403 0.0580 0.0230 0.0292 1.6410 l.0151 0.0101 0.0124 0.0016 0.0017 0.0058 0.0068 1.2458 0.0009 0.0010 0.7847 1.9569 3.6324 0.7045 1.0983 0.1761 0.2249 1.1182 1.6674 I I I 50 6 !I 10 2 20 0.0282 0.0331 50 0.l,026 0.5663 I \.._. I 6 10 20 Note: 0.4056 0.5393 0.0649 o The first entry for E(e - similarly for E(e p _e)3. rcrn:lrks apply to Tnbl~ 3. -- --- 4.5')70 6.6860 1.6225 2.1100 0.1136 0.1164 50 0.1006 0.1231 0.0161 0.0195 ---~.~.. 7.8872 14.7205 2.8394 4.0975 0.7099 0.9355 4 I O/;~, I 0.7721 I l' 0.4613 • X [E(6 p - E(ep»~l/i 1.5994 1.3937 1.0644 1.0629 0.7660 0.7701 0.4533 0.4571 1.3500 1.2636 1.1910 1.6215 1.0212 1.0706 1.0570 1.1666 0.7135 0.7473 0.7418 0.4454 0.4442 0.7641 .. _.~.- 0.4711 ._.' --.-.~--_ ...._*-..._.,- - .__ .. .,-- ..- - 1.9999 1.1904 1.4065 1.2958 1.5670 0.9741 1.0480 1.0558 1.1921 0.7682 0.7500 0.7577 0.6760 0.38LO 0.4259 0.4283 e)3 is the serie::. result to order n-- 2 e - and the second gives the simulated value; ar~ calculated from In both of t.h('aSf~ tR~)les r i~ fixc-cl and s~t 1:'C'!u.al ( . E(e p - E(6 p»3 1.9678 The quantities in the other fou= columns e . I ~ 20 E(6 - E(6»3 [E(8 - E(e»2] 2 % P 6 I I E(e - e)3 • p 3 lE(e - e)2} E(6 - e)3 E(e - e)3 E(e - e)3 p [E(6 - e)~'j% (n) 0 Moment or Measure - --- Sample Size . Value of e .- .... --.__ t() Lhe simulations. t.htJ" s:Hr:n~,c gi~e: in 0:'1("'-, and Similar r.nfd~ .. e f 0( e e e Table 3 Vah~e~ I II Sa~p1e I I Value of 5lZe e I I Cn) ! ii i, j O! : ! I I 1· , J ! I I __ ! 20 ,. I I I , 2 0.OO~8 0.00 ... 8 ~ I I ::>0 0.0008 0.0010 : --_.. ! 2.7648 15.9551 j 6 i I" 1 o 0.9953 3.5496 I , Ito I I · I I I I I !.•- - 6- i I 10 II 4 I 20 1.6326 3.3721 I 50 0.2612 0.3317 I !• 0.007::>, 0.00l:~ i [ECe _,)2] 2 l p . ' "O", .. I I 3°8" ... \ ) : I . I I ~ ! 7.0462 1 ') 9764 ! !I 0 ! ! 3.6149 1 I \ f 7.5414 I i i 4.6589 I I 3.4623 II : 5.9452 I i', 1.6326 2.6946 I 0.2612 0.3004 i • --- 5.8546 I I III I ! 5.1085 ! I 4.1185 I 3.4570 i I i I lI 5.2095 Il 4.1225 I 3.4639 3.4465 6.3697 5.3335 -l 6.2079 I I 1~.1740 5.9784 5.3144 3.4305 1._ f 6.7220 II 5.3458 I i i I !' )I ; I . ! ! I I i i 4.1572 '1: I 3.4335 I J I \ I 6.5551 t I .J ! i i I I I II 5.4503 I I r I I 1. : ij I \I I i I 'orI I I t 4.1292 I I I I I - '-'-'-" . I _.-- ! A98: ! I 8.1311 5.2636 5.3246 ! I i i ' " i ! -• Q! I 14.9475 6.5305 3.4664 I !' '0.248.... ' ' 9 6 6 '; u.• -'"5"0 - ' ,I 'I' a.. ): I . ' . : I. • • • • . • _- I ~• L I I 8.3975 I _.~ I,9.1698 2.7H8 i i \! 0.99)3 - :1 !I 4.1246 I I 0.0008 ! 3.6482 i 0 .0009! I p I 7.8373 ! I I[ECe-E(e))']2 , [E ce -E C'p)) 2J 2 i l I ' I 4.5612 I i 18.1403! 18.1403 107.977( ! 63.5952 6.5305 21.4J.85 I II ' p I " . ECe p -ECe p ))4 I E(e-E(e))~ 'I 0.0192 5.7844 !I 5.1412 ,0.0452 i i i =-=."",,,--==~-=,===",,' i : 9.1721 -- I _,)4 I 0.0398 ! .0-, 0.0550! 0.0/.92 50 =i. !'"'.--=== Ecep -e) II 0.0533 0.2602 C' E 8 E(e~e)l+· A I I[E(e-,-)2] 2 ,~-l+ A I I t 0.2488 .50 72 I J Moment or Measure ' i ' i..., , 0 ,!' 'j' I I' i I , · 0.0192 0.0671 I t · j Ii I I• I ! 0.4250 ! 10 I 1 I J o •0 533 i 6 i ! I Ece-e)' I I I I A. ' I I ! . of the Fourth !loment and MeasurE s of Kurtosis 4.3190 3 3029 ! 4.2632 ! !! I 3.3233 ' I, --'--- 4.'30;j5 3.3300 Table 4 Expected Values .._---_. and Standard Errors of.. _---._ Estimates ----------_ __ ..-.•.-_._---_ .. of Variance for .... _----~-~_._-- Selected Values of _.-.-._--- e and n. ----,.---Moment .~---_._._._---_._._-- Value Sample _o.J.__ .!!__ 2}3.e (n) 10 o 20 40 ..'-t..=r :;:.--=-- 2 =::...-..... -::::=..:=:= E{V2) ._-----_._-1- 0.0948 0.0971(0.0553) 0.0844 0.0849(0.0502) 0.1116 0.1173(0.1300) 0.0437 0.0439(0.0165) 0.0209 0.0209(0.0055) 0.0411 0.0411(0.0157) 0.0203 0.0203(0.0053) 0.0476 0.0479(0.0303) 0.0219 0.0220(0.0096) =.:......=..-:::.=~-=:.~.=;:=-..-...="_-- 0.6660 0.6728(0.3500) 0.6049 0.6009(0.3180) 20 0.3105 0.3116(0.1070) 0.2952 0.2949(0.1010) 40 0.1496 0.1500(0.0354) 0.1458 0.1460(0.0345) ~~~.::l.-::::-:';::' .. ~-:::;:;;:.;. _~-=-'"':'':::':'''''':_-':::=.-=:''::::'=-~;;:'''-=.::j:::=':-:' =:::---:...~=.~~~-::.~ 4 10 1.700l. 1.5492 20 0.7940 0.7562 40 0.3041 0.29'>0 1 --_.•.._- -----Note: ..... I I I 0.7684 0.8074(0.8280) I 0.3345 0.3396(0.2090) j 0.1561(0.0599) 0.1554 ~--:.- . .::.:.:..."""'::--:~~~~.;;..-::..:;:::;=::.z.;.:--:=_~: 1.9567 J __ ._--------_._-_._._._-_._--- . I' ......:.=...~~::-~:::=.=,;..--:-....:: :~..=;-: =.:-.:. 'F'':'=::':'''==--=::'~-_''--'':'::=-~:':''::::'1 10 . E(SP 0.8540 0.3133 ._-------- For 0 = 0, 2 the first entrv in each cell gives the series result to order n- 2 The ~econd is the simulated value and the standard error of the estimate of vari 81l~e is given in parenthesis. Only the seri es results ar.e eben when o = 4. We set r = n in all cases. Table 5 Method I Value Sample of e Size (n) -===--- 90% level level .. "S:'- 10 0 '80% 20 40 ' .. 95.33 80.35 90.32 95.05 (10.47) (5.74 ) (3.48) 90.16 95.21 Note: 90% 95% 80% 90% 95% level level level level level level .._c 76.53 86.99 91.93 81.38 90.28 94.65 78.09 88.35 93.82 80.42 89.99 94.74 (15.11) (8.93) (5.24) (12.84) (7.09) (4.18) 79.47 89.26 94.16 90.04 94.83 I -== ". -- 80.89 _._- ..!"_--- '. 10 80.97 riO.~7 95.25 77.43 87.04 92.14 81.28 90.19 94.65 20 79.40 89.77 94.82 78.14 87.87 93.49 80.06 (11. 39) (6.35) (3.68) (15.10) (9.38) (5.44) (13.57) 89.52 (7.66) 94.43 (4.34) 90.21 95.12 80.15 89.62 94.45 90.0!1 94.95 40 - f-Iethod III - 80% .- -- ::e:==:•• 91.05 ,- 2 95% level 81.17 80.52 Hethod II ,-- 80.63 I 80.78 The first entry in each cell gives the percentage of intervals covering the true parameter value. The second figure given only for n = 20 is the percentage, of intervals that fell totally below that value and so is a guide to the symmetry of the technique. . Ta;,le 6 .~i~_~ r- I I I. Value or w in ~~ estima_te~_ c~ ...P-.ior:. selected vt.lues of nand ¢ --(1- ---- ~ II ' j ord~- _,t~n) _~ 0.5 I !i 10 I 20 6 t.! IIII ! 1 ·'f , 0.1123 0.0939 It i I 6 !I \20 ,_ ----l 50 2.0 I ! 6 0.1667 '-0.1782 0.0900 -0.041':; -0.0055 I~, Exact Bias . l.n ¢* order n-· I -0.0500 -0.0419 -0.0167 -0.0015 -0.0039 -0.0000 0.0412 0.0150 ~ 0.0156. --0.0007 II -0.0006==1 -0.0000. 0.3333 0.5000 -0.5959 II -0.2000 -0.2632 I 1: 0.2150 ij !i 0.2000 0.2600 -0.2526 II -0.0667 -0.0444 II 0.1195 0.1000 0.1150 -0.0299 -0.0158 -o.0002! 0.0424 -O.O_~30 -O.OOz4 -0.00~ 1.0000 1.8333 -1.5837. -1.0000 0.6000 0.9000 -1.1133 -0.3333 0.3000 0.3750 -0.2770 ~ 0.0298 Ii I. I I I -l O.~.::: \. 0.04~0 11--0·5784 20 II 'I 50 ;1 0.1343 10 I 0.0750 I-- . I! 0.0419 ;j I 10 1•• 0 . order n- I 0.0375 !l II I. I. 0.1250 II' _.+"50~_~~:~~~~ • I I. - .. I , ~ ~ Bias in ¢ Bias in ¢ " p ; It-I-----r-------,------j--------r-----ISamnle,' C'" E_X8ct sen. es to1 Serl.es to2 10' Serl .... l.ze 1.il1 l:.xac t 'l 0.1329 0.3688 e .... I 1 l _ I I I I -0.0789 -1.01~ 1 -0.4355 I -0.0287 ~~~_u~-==-~_, -0~~~_-O.0122_ _ '__~~_:~OOO , e e .-: # e e e , Tahle 7 Variance and mean squared error of " ~,¢ p and ~ * for selected values of $ and n. .-r-- Second Order Homents of Value of Sample Size Exact order n (n) Series to order n -2 f j ~ - - - - - - - - - - - p., Variance and M.S.E. Exact Series to . of 9* -2 order n I ! 0.2~50 - 0.1259 0.1276 0.0749 0.0767 0.1250 - 0.1213 0.1213 0.0703 0.071~7.0564 I 0.0564 0.0592 - 0.0595 0.0595 0.0225 - 0.0..24.8. 0.0228 0.0230 - 0.0230 0.0230 0.4130 0.4139 0.6667 - 1.5556 1.6667 .0.0223 0.3114 0.9333 - 0.1673 0.2366 10 0.7115 0.1577 0.4000 - 0.1200 0.7600 0.1015 0.1653 0.4889 - 0.3235 0.3254 20 0.3362 0.3505 - 0.2000 - 0.2800 0.2900 0.1800 0.1809 0.2211 - 0.0800 0.0928 0.0823 0.0833 0.0823..., 0.0154 5.4000 2.5235 - 0.0835 0.0835 0.1427 1.1725 0.1200 1.3595 2.6000 - 0.4850 0.6741 0.1875 - 0.3438 0.3594 10 0.2116 0.2204 0.1125 - 0.1688 0.1744 20 0.0752 0.0769 0.0562 - t-~~-,,_0.?52 10.0250 6 50 1.0 50 10.0953 0~.()3.5~ I I I I ~.O~~_~ I 0.2241 0.2241 - q~O?~~ 0.1143 3.0000 - 10.0000 11.0000 10 1.2711 1.2887 1.8000 - 4.0800 20 1.8908 2.0268 0.9000 - 1.5300 1.6200 0.3872 0.~639 1.0895 - 0.9161 0.9169 50 0.5013 0.5193 0.3600 - 0.4608 0.4752 0.3708 0.3111 0.3894 - 0.3939 0.3939 I +.<?·~2,.Zl 6 2.0 Note: Variance & M.SoE. of 0.0322 0.0640 0.2783 0.2909 I I -1 6 0.5 I Series to 4> 0.37~7 4.~200 I The first entry in each cell is the variance and the second the mean squared error. , Table 8 Third and fourth moments of ¢ and ¢ and measures of skewness and _______________.__. .__ __..-..P_._. . . kurtosis fva1ue I ~_. I ~" C ... I· If' .- 0.5 ! E~(jl_¢;)3 S~~e ! I .(n! L I 6 II I 10 20 I' - 6 I 10 . 1 0 I . , I! +- 20 I i I 2.0 0.2956! 0.0073! 2.8569 ! ! 0.0396 0.0150 t I I E(<j>_¢)4 I i 1.8565 I I 0.2923 ! 0.7108 0.1588 ; ! 1.1225 ! l 0.7215 I -1.0881 0.0515 II . 10.0028' 0.5965 i 2.1979 0.0710! 2.7497 -0.0889 I 0.5706 I 0,~41_4 6 i-0.9313 I I 10 1.3267 I 20 7.7239 I\ I ! ! I; -1.3235 \ I ! 0.9225 \ 0.3403 I ! ! 0.0182 I ! 14.6330 !I' 8.7001 I , 0.0168 I 0.0021 l I -4.0446 I I I -1.5425 l !i -1.7931 -0.2805 0.75~_ ~2563J e ~. 0.9068 2.6767 2.0093 I Ii i· I ; i I I! 2.0473 II 0.1493 Ii I I r 4.0537 4.5605 16.6637 I ! 5.5321 : 4.7259 6.5184 2.5096 'I 1.0236 , 1.3562 l 2.1508 I -0.887643.9895 I 0.4630 I'10.7080 ~.8147 , I I 2.1742 2.4109 e 5.7145 I: 2.5066 1.1337 ! 'I' ! l~ I 1.2411 l 4.0043 -1.1312 2.8519 i ! I! I ! 1.9866 ,'7.0053 ! " II --- 0.0594 \1.2806 2.3053 I 0.1768 ..---.-.- I I _ 4.4301 4.0220 I '-1.0089 .~ i I ' : JI I I I ! -·-A---- . {E(,;._<p)2}2\ {E(<jl _¢)2}2 I I ~~~1_0~~?8_=t-_~~~_!~~1_ 0:,0.:'21 b-<l.·~320 ' ! oC. • E(ip-<jl) 4 I, ---;;-----, I i ! ' I -0.2523 ' E(¢-¢)4' 11 I ---il. Y I O.OC~~ 5.2702 I I I I 0.3454 . 4 E(q> -q» p ·-:-·,:-::::.-::=--;::'f::=-=-":"'·-:'·:-::-':::·::-=·==;:=:::=:::·:':==-=:::--":":.F --- I 1.4497 I I -1.4223 I I I i I I j : 1 0.0042 1.0445 ·-1==::=::·-·-::::::::::'"f====::::-·:.:::;-·-:::;:.-.:::-.:.:;:::::::::···-_· .. ! I {E(<j> _¢)2}12 l , i ; i 0.0025! ,50: L -0.0230 I ---~ - .--:3:/ . +--!---i--· 1.8633 50. i I{E(9-<P)2}~Y2 I I I I I . I I I -A---~- ! I 50 _ 11 P _<jl)3 l I II E(</! .----l------r ! i I 1 0.2923 I i ~--~-II--:(~-~~II--E~~p-~)3 -l~--A---·----~. I .' I I , I I. ~~le [9.8284 I 5.9143 e I Table 9 Estimates of Variance and Hean Squared Error of cjJ and <p • ·-·:·-::'::-~::'=':"-===-===_::=:==--:-":··-::==··'::::::=:~:===:-::=-""'=-':;==E...---"A-·_...... Value of ¢> Homent Samp Ie I - - -..---r--------,~~---___l :ni;e - 0.5 E(Vl) E(V2) E(Si) V(cP) and V(</J p ) and A H.S.E.(</J) p i-.-.----+-----------I-----..:.=----+------\-------t 0.2544 (0.6569) 0.1244 (0.1279) 0.3647 (1.6963) 0.2116 0.2204 0.0749 0.0767 20 0.0795 (0.0975) 0.0658 (0.0642) 0.8281 (0.1372) 0.0752 0.0769 0.056/. 0.0564 50 0.0255 (0.0143) 0.0241 (b.0131) 0.0256 (0.0146) 0.0250 0.0252 0.0228 0.0228 _ _. '::7_7=-"'=-;-":-"_-0.."1=::::._ 20 A H.S.E. (</J ) 10 10 -- 1.0958 (2.0659) 0.3639 (0.8744) ...... 0.2717 (0.1971) 0.2368 (0.2492) I - - 2.2167 (5.9471) 0.4552 (2.3209) 0.7115 0.7577 I 0.1015 0.1653 I 0.3362 0.3505 0.1800 0.1809 0.0953. 0.0971 1.2711 1.2887 . J 2 0 2 0 2.4842 (5.7664) 0.7807 (0.6078) 5,0032 (18.6861) 1.8908 2.0268 0.3872 0.4639 0.5145 (0.6498) 0.4171 (0.3822) 0.5701 (1.1518) 0.5013 0.5193 r,.3708 50 _ _ _ _ _ _ _• Note: 1-.. _ _._. .. _ _._ _. . • .. _ .._. u.3711 ..__ -----_._- .,I. The first entry for the last t\110 columns is the variance of the estimat.:.. and the second the mean squared error. The figure in A parenthesis for the E(V ) and E(Si) columns is the standard error i of the estimates of \7ariance. . Table 10 ._---.-----_._ _._..__... _--... Value of Sample Size (n) _..-80% level Method I 90% 1 eve1 Nethod II Nethod III 8-6% ·-'9"0"%-·--9·5·7...·- 95%level level level level . _ - - - . - 10 91.17 8.83 LOS 9 1.17 8.83 1.34 91.17 71.31 8.83' 28.69 1.60 0.82 91.17 8.83 1.05 81.11 15.13 0.66 84.87 15.13 93.98 6.02 .83.58 15.13 78.61 17.15 0.40 88.56 10.35 0.51 _. 94.09 5.70 0.61 - 78.61 17.15 0.39 8 2.71 1 7.29 2.71 94.58 178.26 5.43 21.74 3.23 1.24 ---'-- ,. 0.5 20 50 . r.~n I 10 17.29 .~ 2.11 - - _. . . . . --------~. 0 _ _' - •• -80% 0 •• _ 28~69 1.21 91.17 8,83 .1.60 84.87 15.13 84.87 15.13 84.87 15.13 84.89 15.13 88.56 10.35 0.49 94.21 5.70 0.59 78.61 17.15 0.40 78.26 21.74 1.60 78.26 21.71. 1.90 82.71 17.29 2.82 ~ - . _ o 71.31 _ 91.17 8.83 1.25 __ .__ ~_~.Q_~L~ __.__~_2}__Q.:!.? -.;;:-...=.:~- y 74.7~·~;--86~2 94~23 level • g_L~?._._oJ:~oQ..!. ! 90% level 91.17 8.83 1.98 ~ I 93.98 6.02 0 • 99__ 1. O~ 89.16 94.27 i 10.35 5.70 Jl 0.52 0.62 ------_. 82.71 17.29 3.73 I 94.57 5.43 4.60. I 86.84 86.84 94.23 1 13.16 13.16 5.77 1.46 1.90 2.30. I 82.25 89.85 94.05 l 16.11 10.13 5.95 I 0.79 1.01 1~ - - - - -..:;==.-"-";=----------= ------- ==:.:.70==. '=--";"--:;:=--";:=-="::::-';;"-==---=-"=..:.---=-=-===-===_==1 i I 1.0 ~ 184.79 86.84 20! 13.16 13.60 5.77 25.19 13.18 , 1.33 1.71. 2.03 1.14 1.46 to----j----7-7-.-9-4--89-.-1-0 ~-'-~-;--80.6-;: 89.10 50 16.11 10.13 5.95 16.11 10.13 0.76 0.98 1.17 0.73 0.94 Note: i 10 76.21 23.79 3.93 91.4 6 8.5 4 5.0 4 91.46 8.54 6.01 29.03 70.97 1.46 54.44 45.56 1.88 54.1.4 45.56 2.24 76.21 23.79 6.06 91.46 8.54 8.04 91.46 8.54 9.92 20 80.88 19.12 3.18 90.7 8 9.2 2 l~ .0 8 90.78 9.22 4.86 79.45 20.55 2.10 79.45 20.55 2.69 0- 89.35 10.65 3.21 80.88 19.12 4.01 90.78 9.22 5.22 9C.78 9.22 6.32 50 84.56 12.59 1. 70 92.2 7 7.5 6 2.1 8 7.56 2.60 19.64 1.55 87.36 12.59 1.99 92.44 7.56 2.37 80.36 19.64 1. 79 87.41 12.59 2.31 92.44 7.56 2.77 I l I 13.18 1.74 94.04 5.95 1.12 - ----------- n'4j17.51 - _.. ---- ._-~-- ) The first entry in each cell is the actual probability uf the interval covering the true value of ¢ and the second is the probability that it will fall completely below ¢. The third entry is the expected length of the interval. Table 11 ,.. Bias in ~ "and ~ with -------_.-.......,....P -- va-1u~"--I-----~~~~' in ~-~-=-~.of A ~o order n- 1 ---r-TO order n- 2 ---------1----- - --.......----- n =5 "-~--_._--_. "-----~----i Bias in ~ Bias in A as i P p to orqer n- 2 a percenta~e of bias in A. --~ .. _~-- -- ----- 0.5 -0.0239 -0.0223 -0.0020 9.1 1.0 -0.0345 -0.0336 -0.0012 3.5 1.5 -0.0374 -0.0313 -0.0001 0.2 2.0 ...0.0359 -0.0365 0.0007 2.0 2.5 -0.0322 ...0.0332 0.0012 3.6 3.0 -0.0276 -0.028R 0.0014 4.9 3.5 -0.0229 -0.0241 0.0015 6.2 4.0 -0.0184 -0.0196 0.0015 ------- . ... I __ 7'5._J . .. Table 12 --Second Order Moments of A and Ap for Selected Values of A with n • 5 ~ Value of A II I I I - ~ V (A) and E (A-A)2 to various orders se se V (~)=E(~ -A)2 se p p A Vse(A) and E (A-A)2 to n se -1 -2 A Vse.(A) to n to order n E (A-A)2 to n -2 se ~ A Vse (\) " Vse(A) -2 to n -2 E ~e (A -A)2 p 1 . - - - A - E (A-A)2 se -2 to n nLE se (A-A)2 0.1 0.0387 0.0343 0.0343 0.0388 1.132 1.131 1.128 0.3 0.1091 0.1001 0.1004 0.1098 1.097 1.094 1.087 0.5 0.1716 0.1617 0.1623 0.1731 1.034 1.032 1.057 1.0 0.3024 0.2976 0.2988 0.3054 1.013 1.011 1.5 0.4095 0.4118 0.4132 0.4130 1.001 1.000 0.991 2.0 0.5035 0.5117 0.5130 0.5067 0.995 0.994 0.981 2.5 0.5911 0.6034 0.6044 0.5937 0.992 0.991 0.978 3.0 0.6765 0.6909 0.6917 0.6784 0.9:H 0.990 0.978 3.5 0.7619 0.7772 0.7777 0.7632 0.991 0.991 0.980 4.0 0.8487 0.8638 0.8641 0.8495 0.992 0.991 0.982 . , II 1.012 i e -, • e <: , e Figure 1 Bias in the three estimates A A ~, ~p and ~* Bias 0.8 Series to order n.- l for ¢ 0.6 0.4 0.2 10 20 30 40 50 0.0 t----t.:.:---~;::===.....;.;...-_=~::===:---+ -0.2 -0.4 -0.6 -0.8 -1.0 Sample Size • Figure 2 The variance of $, $p and ¢* • Variance • 2.0 1.6 . 1.2 0.8 A </> </>* 0.4 10 Sample Size Figure 3 Mean squared error of the three estimates Mean squared I error 2.4 • 2.0· 1.6 1.2 . 0.8 0.4 10 20 30 40 5 Sample Size
© Copyright 2024 Paperzz