Bell, Walter E. and VanDerVaart, H.R.; (1972)A method of estimating the coefficients in differential equations from time discrete observations." Thesis.

•
...
•
A METHOD OF ESTIMATING
THE COEFFICIENTS IN DIFFERENTIAL EQUATIONS
FROM TIME-DISCRETE OBSERVATIONS
by
Walter E. Bell and H. R. van der Vaart
Institute of Statistics
Mimeograph Series No. 845
September 1972
•
•
•
iv
TABLE OF CONTENTS
Page
........................
1.
INTRODUCTION
2.
REVIEW OF LITERATURE
1
4
2.1 Estimation of Coefficients of Differential
3.
10
THE SOBOLEV NORM-TYPE EXPRESSION, SMOOTHING AND
DERIVATIVE ESTIMATION, AND ESTIMATOR COMPARISON
13
3.1
Considerations Regarding Use of a
Sobolev Norm-type Expression • • • • • • •
5-point Moving-Arc Polynomial Smoothing and
Derivative Estimation • • • • • • • • •
3.2.1 Smoothing and Derivative Estimation
Independent of Estimation of K. .
3.2.2 Parameter Estimation with Simultaneous
Estimation of K • • • • • • • •
3.3 A Function of Segmented Cubic Polynomials
3.3.1 Least-Squares Spline Functions
3.3.2 Construction of the Segmented Cubic
Polynomials • • •
• • • •
3.4 Comparison of Estimators • • • • • •
•
4.
Genera~
ESTIMATING
4.1
4.2
K
IN
.9-Z
=
dt
•
13
16
16
20
21
22
25
31
35
Ky
Estimation of K and Smoothing Operations
Conducted Separately • ••
••••
4.1.1 General Development
4.1.2 Numerical Example •
Simultaneous Estimation of K and the Parameters
of the Smoothing Function • • • •
4.2.1 General Theory
4.2.2 Numerical Examples
..
..
4
Equations
• • • • • • •
• • • •
Investigation of Observational Error Effect
2.2
36
36
45
51
51
55
•
v
~BLE
OF CONTENTS (continued)
.
Page
5.
GENERALIZATIONS OF THE SIMPLE CASE
5.1
Estimation of
5.1.1
5.1.2
5.2
5.3
6.
.
7.
...
.
•
K
1
and
K
2
in
.=
y
62
K Y + K
1
2
z
Estimation of K Following Initial 5-point
Polynomial Smoothing and Derivative
Estimation
• • • • • • • • •
E~timation of
K and K with Simultaneous
1
63
63
2
Smoothing •
• • • • • • • • •
Generalization to a System of Differential Equations
Generalization to a System of Differential Equations
in Which Some of the Coefficients Are Related or
Assume Known Values • • • • • • • •
• • • •
66
68
77
BIAS REDUCTION
83
6.1
6.2
83
General Considerations •
Examples of Reducing Biastotal
85
SUMMARY AND OVERVIEW OF OPEN PROBLEMS •
7.1
7.2
Summary • • • • • • • • • •
Overview of Open Problems
8.
LIST OF REFERENCES
112
9.
APPENDICES
116
9.1
9.2
116
118
Derivation of Equation (4.11)
Derivation of Equation (5.26)
•
1.
INTRODUCTION
Differential equations of the form
dYj _
dt
-
k
t K.. Yl.'
j
i=l Jl
= 1,
••• , L, 1
~
L~ k
(1.1)
are frequently used in the mathematical models associated with biological
systems.
For example, differential equations of this type often appear
in the models of drug disposition in the human body.
Certain models of
population dynamics also employ differential equations related to
(1.1).
Assuming the relevance of (1.1) to the true underlying biological
phenomena, interest has been traditionally focused on estimates of the
values of the
K..
Jl
For observed values
in most experimental applications of this model.
y.(t ) = y.(t ) + €.(t ) ,
1
m
1
m
1
m
where
E.
1
(t)
m
is a
K.. has been
Jl
accomplished in most of the existing literature by a two-step procedure:
random error and
m = 1, 2, ••• , n,
estimation of the
~
(1)
A solution function
y.(t)
or numerical methods.
for the
t = t
(2). Values of the
~
m=l
of (1.1) is derived by analytical
J
Such a function is defined at least
m
K..
Jl
are chosen to minimize the quantity
[y. (t ) - ; ; ; )J
J
m
J
m
2
",........
Y. (t ) denotes the value of
J m
of observations.
where
..
•
(1.2)
~
y.(t )
J m
for a given set
Estimates of the
K.. derived from minimization of (1.2) are usually
Jl
referred to as "least-squares" estimates.
Except in conjunction with certain enzyme kinetic models (Cleland,
1967), apparently little work has been done on estimating the
K••
Jl
•
.
2
The few examples of existing
without solving the differential equation.
methods appearing in the literature involve estimating the value of the
dye
derivative
it-
t
at the
m
from eXPerimental observations by such
methods as difference schemes (~.~., see Rescigno and Segre
and, by choosing the
IC . .
J~
n{·[k
I:
I:
m=1 .
. 1
~=
p.
9»,
to minimize
/'-..}2
IC .• Y.(t)] - y.(t ) .
J~
~
m
m
J
dy.
denotes the estimates of ~
where
(1961,
I
,
t=t
estimating the
m
IC..
J~
in a manner analogous to that employed in linear least-squares
approaches to multiple regression.
~I
dt
values
t=t
However, the estimation of the
from the discrete observations
m
absence of an analytical solution y.(t)
J
considerable care in the choice of
(1966),
method~
Y. (t )
J m
in the
has traditionally required
(See, for example, Joksch
who discusses the effects of random observational errors on the
derivative estimates derived from the least-squares fitting of an
approximation function to the observed values.)
The purpose of this thesis is to formalize a method of estimating
the
IC..
J~
in (1.1) which is based on minimization of a quantity, the
expression for which is related to a discrete version of the Sobolev
norm.
For several differential equations of simple form, some of the
properties of the derived estimators
..
•
IC •.
J~
are investigated by both
approximate analytical and Monte Carlo simulation methods.
To insure
relevance to investigations of mathematical models related to biological
phenomena, the number, n, of observations is limited and the distributions
•
·
3
associated with the observational errors,
E,
are chosen to permit study
under both constant variance and variance proportional to the
y.(t ).
m
l
A review of some of the literature involving estimation of the
coefficients of differential equations of particular biological
Chapter 3 contains definitions
significance is the subject of Chapter 2.
and construction methods for 5-point moving-arc polynomial smoothing and
derivative estimation and for a particular type of spline function
composed of cubic segments.
The Sobolev norm-type expression and
several parameters useful for comparing estimators are also introduced
in Chapter 3.
In Chapter
4, the estimator of
K
•
in the simple
differential equation
•
.<?1:
dt
is derived and investigated.
= K
1
Extensions of the method derived in
Chapter 4 include estimation of
K
1
~-K
dt - 1
K..
and the estimation of the
.<?1:
=
dt
dz
and
y+K
a
K
a
in
z
in the system
Jl
dt
Y
K
=K
11
21
Y + K
12
y+K
2a
z
{1.4)
z,
where there may be, but not necessarily be, restrictions on the
K.
l
•
These extensions and examples of their use constitute Chapters 5 and 6.
•
•
4
•
REVIEW 0 F LITERATORE
2.
Estimation of Coefficients of Differential Equations
2.1
Nearly all of the current methods of estimating the coefficients
of differential equations rely on either a knowledge of the form of the
analytic solution or on the existence of the values of a numerically
generated solution at the points of interest,
corresponding to experimental observations.
the points
~.~.,
In fact, in the area of
tracer kinetics, the methods of estimating the coefficients are
generally described as the estimation of the
A.1
("rate constants")
in the expression
m
•
=
f(t)
I: A. exp ( - A. t ) •
i=l 1
1
For homogeneous ordinary linear differential equations with
constant coefficients (whose solutions are essentially sums of
exponentials), the coefficients of the differential equations appear
as the
A.
1
or some function of these coefficients appear as the
In such cases, several methods of estimating the
A.
1
A. •
1
are available.
The simpl.est is the common "peeling-off" technique described in detail
by Perl
(1960).
tecb~ique
Cook and Taylor
(1971) describe a computerized peeling
for application to radioactive tracer efflux data and report
the results of simluation studies for two- and three-compartment
systems for different
A.
1
and various relative errors.
peeling process assumes that the
•
•
Ao
1
Since the
are sufficiently different to
provide a linear terminal segment, the observations must be carried out
far enough in time, which can be experimentally difficult •
•
5
a
...
For these models of radioactive tracer dynamics, Gardner {1963)
proposes a Fourier transform method which is described by Pizer et ale
(1969).
Use of this method requires interpolation and extrapolation
of the data to accommodate the integration schemes employed.
For
application to biological experiments with their large errors and
short duration, this method may fail.
other methods of estimating coefficients generally involve seeking
values of the coefficients which, for a given set of data, minimize
the sums of the squares of deviations
SS =
nt (y. _ y.)2
"
i=l l
l
•
where the
y.
l
are the observations and the
the solution (analytic or numerical).
are the values of
While not often specifically
stated, the analog computer procedures generate curves for a given
set of parameter values; each generated curve is compared with the
data points in an attempt to determine the "best fit", resulting
essentially in a minimization effort applied to the sums of squares.
The coefficients
K..
lJ
in the differential equation, say
dq.
k
---.=s:L - "
~ K . .q.
dt
lJ l
i=l
which is common in tracer kinetics, are usually represented by
variable potentiometers on the analog computer.
Varying the
K:.. ,
lJ
one attempts to generate a curve which fits the observations more
•
closely.
Heinmets (1970) discusses in detail and expounds the virtues
of the use of analog computers in biological model evaluation.
•
6
·
...
Analog computer notation and symbolism are used in such package
programs as the IBM Continuous System Model Program, which is used
by Parrish and Saila
(1970) to estimate the coefficients of a system
of Latka-Volterra competition-type differential equations aod a
modified Volterra predator-prey equation.
Evert and Randall
(1970),
encourage the use of the matrix approach to systems of differential
equations in the context of tracer kinetics and cite the value of the
Continuous System Model Program.
When an analytic solution to a differential equation (or a
system of differential equations) is not available or when the solution
is not linear in the parameters to be estimated, then a non-linear
•
regression scheme is frequently employed to determine estimates of the
parameters which minimizes the sums of squares of deviations.
Rosenbrock and Storey
(1966) and Swann (1969) present reviews of
current methods, most of which are iterative and employ a linear
approximation at every step.
Nearly all such procedures require
initial estimates of the parameters and can experience difficulty in
convergence.
Further, in the case of differential equations for which
analytic solutions are not available, numerical integration procedures
are a necessary part of the regression technique.
Cannon and Filmer
(1967) describe in rigorous detail some of the
mathematical properties of the estimates of the rate constants found
in chemical kinetic models.
(Cannon and Filmer,
•
The authors provide a numerical example
1968) in which the initial conditions and the
error rate are varied to demonstrate the behavior of their estimation
•
..
7
method.
Their original system of simultaneous differential equations
is solved by finite differences while their norm
is evaluated by the trapezoidal integration procedure.
Procedures for estimating the coefficients of (a set of) ordinary
differential equations without use of an integration scheme are
apparently rare, with the exception of an important class of enzyme
kinetic models.
Since most enzyme kinetic experiments are performed
by measuring initial reaction velocities at various initial substrate
concentrations, the estimation of the rate constants
K.
in the
l.
differential equation
...,.
K , K ,
1
:3
... )
may be accomplished by a least squares scheme if the velocity equation
is linear in the unknown rate constants or by a nonlinear technique if
the unknown rate constants appear nonlinearly.
"observed" values of
v
= ~;
are made from either a continuous
recording device or by fitting a function
the discrete observations on
In either case, the
(~.~.,
a parabola) through
P, of which there are a large number
(Cleland, 1967).
Whittle (1956) considers a time-dependent infection rate function,
A(t) , of wild rabbits in the integral equation
t
R( t) =
•
J
o
where
R(t)
t
exp
( J [A(V)
- ~(v)J} A(u)du
u
is the ratio of infected to healthy rabbits and
the death rate due to infection and assumed to be 0.45.
~(t)
is
He notes that
•
.
8
the above integral expression is a solution to the differential equation
=
dR(t) + [e(t) - A(t)] R(t)
dt
He proposes estimating
R{t.) ,
dR
dt
and estimating
J.
Metzler et ale
by difference quotients, eye-smoothing the
A(t)
A(t)
A(t) •
at discrete points by
= R' ( t)
+ @(t) R ( t )
1 + R(t)
(1965), in their work with non-steady state tracer
kinetics, propose a similar approach to estimating the transfer rate
function
B(t)
of sodium from the plasma to rumen of sheep in the
differential equation
dA
dt
:a
B(t) [A {t) - A:a(t)]
l
=
N (t)
2
where
A
l
and A
:a
are the specific activities of sodium in the plasma
and rumen, respectively, and
rumen.
N
is the concentration of sodium in the
2
The authors suggest that each of the
N , A , and A
2
l
be smoothed
:a
by a moving-arc polynomial scheme, for instance, from which
dA
:a
dt
caL
Then,
be estimated at the times at which observations were performed.
B~t)
can be estimated at the times of observation
t.
J.
by
N (t.) A '(t.)
B (t.) =
J.
•
:a
J.
:a
J.
Al (t.)
- A (t.)
J.
:a J.
In the context of drug kinetics, Martin
of determining
(1967) proposes a method
K, the rate constant for elimination of drug by all
•
9
k
routes, from measurements of unchanged drug in the urine.
Assuming
first-order kinetics, he derives the differential equation
dD
U
dt
where
time
D
K(D
um - Du )
is the cumulative amount of unchanged drug in the urine to
u
t
=
and
Dum
is the total cumulative amount of drug in the urine
after excretion is complete.
distributive phase,
For times
t
after the absorption and
K is estimated by considering the linear relation
which results from substitution of
6D
for the derivative.
dD
u
M
u
estimates
at the midpoint of the urine-collection interval
dt
6D
of
According to Martin, since
u
6t
6t , the values
dD
would have to be interpolated to yield estimates of
the end-points of interval
D
u
The author shows that, when the
dD
d't"""u
at
At, to coincide with the times of
observations of the quantities
decline of
u
~
is first-order, the substitution of
an error which would not exceed two percent even if
6D
u
At
At
would produce
were as large
as one half-life of the drug.
Rescigno and Segre (1961) note that the estimation of the rate
constants
•
where
K
1
and
K
2
in
Q is the constant concentration of a substance in an external
•
-
10
medium, can be accomplished by numerically estimating the derivative
dX(t) at
dt
t=t.
~
by a certain difference scheme and writing the
differential equation in the form
X = M- NX(t) = {l~X) [ MNJ
which is recognized to be a special case of the usual linear expression
y
in which
and
N
eo
Y= X ,
(hence,
"
K
1
= K Q ,
1
and
=
eo
+
eX
1.
e1
and
Then estimates of
= K
a
M
~ ) can be obtained in the customary fashion
a
for linear least squares models.
The potential value of this type of
approach is obvious if fairly accurate estimates of the derivatives can
be obtained.
In fact, Rosenbrock and Story
(1966) remark that, if we
could measure the derivatives directly, then estimation of the rate
constants would not involve the solutions to the differential equations;
however, the authors don't pursue their remark.
2.2
Investigation of Observational Error Effect
Although many authors have dealt with the techniques of estimation
of rate constants, and, hence, of the coefficients or functions of
coefficients of differential equations, relatively few authors have
reported the results of simulation studies in which the performance
of the various estimation schemes was investigated under specific data
error and rate constant magnitudes.
•
In general, these studies involve
simulated data which are constructed from hypothetical situations such
as could be expected in biological applications, where the number of
•
•
11
samples is limited and where the accuracy assumed is not normally found
in non-biological situations •
In the context of tracer kinetics, MYhill (1967) reports on studies
involving the sum of two exponentials with positive coefficients,
-A t
1.
f(t) = N e
1.
with
N , N , A , A >0
1.
:<I
1.
:<I
,
+ N e
:a
!.~.,
-A t
:a
for specified ratios of
N
1.
IN:<I
under the limitation of 11 and 31 equally spaced points.
and
A1
IA :<I
Using a
"valid least squares gaussian iterative" technique, MYhill found that
for 11 points and a ratio of
A
1.
IA :<I
=
2
the technique did not
converge for data sets with more than 1% error.
A similar analysis of
the tracer activity curve of a three-compartment steady-state open
system, which is represented by a sum of three exponentials with
positive coefficients, is reported by MYhill (1969).
In this analysis,
each point was weighted in order to reduce the differences in estimates
between the cases of percentage error and constant error.
In work similar to that of Myhill, Glass and deGarreta (1967)
report on error analyses in which a sum of two exponentials was
fitted to generated data with error using the Marquardt method and a
method based on the Newton-Raphson technique in which the weighting
of data may be incorporated.
Unfortunately, both the Myhill and the
Glass and deGarreta results are unavoidably restricted to monotone
decreasing functions.
Although Westlake (1971) did not conduct an extensive error
•
investigation, his report appears to be the only available which has
discussion on the effect of data error on a two-compartment open model
•
-
12
which is represented by the sum of three exponential terms
•
-A. t
A. e
1
1
in which the
A.
1
and the
A.1
are functions of the rate constants
and the volumes of distribution.
Such sums of exponentials are
distinctly not monotone decreasing in the drug kinetic case in
which the drug is orally administered.
Westlake discusses, by means
of an example, the effect of a constant error in the plasma
concentrations on the estimation of the rate constants (and the
functions of the rate constants) in the solution of the differential
equations for the plasma concentration.
Cook and Taylor (1971) give tabular examples of the performance
of their "peeling" computer routine in the estimation of rate
constants for two- and three-compartment systems in a tracer efflux
system•
•
•
-
13
3.
..
THE SOBOLEV NORM-TYPE EXPRESSION, SMOOTHING
AND DERIVATIVE ESTIMATION, AND ESTIMATOR COMPARISON
3.1 General Considerations Regarding
Use of a Sobolev Norm-type Expression
In this chapter and the next, methods are formulated to estimate
the constant coefficients of a certain class
o~
first-order ordinary
differential equations which are linear in the coefficients.
Although
the techniques may have more general application, the formulation will
be restricted to a rather simple class of differential equations.
Consider the differential equation
where
K is a constant
estimated.
(l.~.,
time-invariant) coefficient to be
Observed values of
by capital letters,
~.~.,
y
at times
t=t.
1.
will be denoted
y(t.) •
1.
Although most current techniques of coefficient estimation use
some scheme which seeks the value of the coefficient
K which
minimizes either the sums of squared deviations of a solution
(analytic or numerical) from the observed values or, as in the
example of Cannon and Filmer (1968), the integral of the squared
deviations, the estimation procedure proposed in this discussion seeks
to minimize the expression
1l =
•
n
2
I: {[y(t.) - f(t.)J
1.
1.=
.11.
. )
+ w[y(t.) - g(t1..)]Z
1.
which is somewhat analogous to the discrete version of the Sobolev
norm.
In equation (3.1),
.
y(t. )
1.
denotes an "observed value" of the
•
14
derivative
%t
points
of observation by means of the model
t.
~
expressed in terms of the observed values of
.
=
y(t.)
~
w
~
and
0
is a constant (usually
and
f(t.)
~
g(t.)
at the
K y(t.) ,
~
w
= 1)
with dimensions of time-squared,
are approximations to
~
y
y(t.)
~
and
.
y(t.)
~
obtained from the data by smoothing and derivative-approximation
procedures.
Then, in the current example,
ut = t.
([y(t.) - f(t.)J2 + w[K y(t.) - g(t.)J2}
~
~
~
~
~
and, not only an estimate of the unknown coefficient
any parameters in the functions
f
so as to minimize the value of
va.
shows that the
of the function
and
g
K, but also
will have to be chosen
However, the model as written
K directly determines the values of the derivative
y, so that it is natural to include in the selection
criteria the minimization of the squared-deviations of derivatives of
y , as well as the values of
Y itself.
Further, the incorporation of the differential equation itself
in the quantity (3.1) to be minimized and the appearance of
K
linearly in the differential equation permit
K
instead of a function of
estimation of
K, which often results when only a
solution curve, in which the
K does not appear alone or linearly,
is considered in the quantity to be minimized.
The functions
•
•
f
and
g
are arbitrary at this point, although
several of their desirable potential properties are obvious:
(1)
f
i
and
th
g
should be defined at each time
observation is made.
t. , when the
~
•
15
(2)
..
f
and
should have smoothing properties, 2:..~., their use
g
should reduce the effect of experimental error •
(3)
f
and
g
should be useful regardless of the "shape" of the
graph of the underlying data function.
(4)
g
(5)
Ideally,
should be reasonably insensitive to experimental error.
where
t
f
1
and
t
~
~
g
t
should be defined for all time,
t,
n
Such properties suggest a number of potential forms of the
functions
f
and
defined at the time
g.
For example,
t., i
1
= 1,
f
could be the point function
2, ••• , n,
which results from
application of a polynomial moving-average smoothing function and
g could be the value of the first derivative with respect to time
of the smoothing polynomial applied to the point
hand,
g
On the other
t.1. .
could be the point function defined at the
t.
1.
which
results from some numerical differentiation technique such as a
central differencing scheme.
Of course,
f
could be the polynomial
defined over the entire time domain of the data, a technique used
by Lewi et ale (1970) to represent radiochemical data.
These
authors recognized the limitations imposed by fitting one polynomial
to the entire set of discrete observations; for instance, a low
degree polynomial may poorly approximate the data while a polynomial
of high degree may better represent complex relationships in the data
at the expense of smoothing and may also introduce excessive variation
in the derivative estimates.
•
One important group of functions which
shares some of the desirable attributes of polynomials, which does
•
16
~
not require the fitting of one polynomial over the entire range of the
data but which is defined over all
t
such that
t
l
~
t
~
t
is a
n
set of spline functions.
In this dissertation
t
f(t)
is restricted to those functions'of
~.~.,
which are linear in their parameters,
defined at
t
(a) the function
which is derived from 5-point moving-arc smoothing
i
using either the cubic polynomial
p(t.)
~
a . + a .t. + a .t. 2 + a .t. 3
=
o~
l~
~
2~
~
2~
(3.2)
~
or the linear-hyperbolic function
q(t.)
~
=a
1
.t. + a . + a . -t
o~
~
l~
2~.
~
and (b) a spline-type function composed of cubic polynomial segments.
Similarly, the function
defined at
t.
g{t)
is restricted to (a) the functions
which are derived from the first derivative with
~
respect to time of the two smoothing functions
p{t)
and
q(t)
and
(b) the derivative of the segmented cubic polynomial.
3.2
5-point Moving-Arc Polynomial
Smoothing and Derivative Estimation
3.2.1
of
Smoothing and Derivative Estimation Independent of Estimation
K
Consider the
of the
•
•
y(t.), j
J
n
observations
= 3, 4, ... ,
y(t.), i = 1, 2, ••• , n.
~
Each
n-2 , can be replaced by the value of
the 5-point smoothing function {cubic or linear-hyperbolic) which is
obtained by the fitting of the smoothing function to the five points
y(t.), i
~
= j-2, j-l, j, j+l, j+2, by the method of linear
•
...
•
17
least-squares.
In the usual notation, the estimates of the
a.
-).
in
equation (3.2) can be computed
A
T
-).
-J -J
a. = (X. X.)
-1
T
X. y.
-J-J
where, for the cubic smoothing function,
J-a
t . at . 3
J+a
J+a
t '+
J a
1
Then, the smoothed estimates of
y*( t .)
=
J
...
3
t.
1
[ 1
y( t.)
J
y( t. )
J+a
can be computed
t.
J
t. 2
J
t. 3 ]
J
( X. TX. ) -1 X. Ty.
-J -J
-J-J
t.
J
t a
J
t
(X. TX .) - l X .T
-J -J
-J
s. Ty.
= -J
-J
where
s.
T
-J
=
[ 1
]
3
J
To obtain arbitrary smoothed values of the first two and last two
observations, the vectors
S
T
-1
•
t
-2
=
[ 1
t
T
s
-n-1
=
[ 1
t
=
[ 1
t
s
...
T
[ 1
s
-n
T
s
-1
T
s
T
1
t a
1
t
a
t a
a
t
n-1
n
t
;c
n-1
t a
n
s
-n-1
-2
t
t
1
and
s T are defined
-n
3
]
(X Tx )-l X T
-3 -3
-3
3
]
(X Tx )-l X T
a
3
n-l
3
n
T
]
-3 -3
]
-3
(X
Tx
)-l X T
-n-;c -n-a
-n-a
T
(X
Tx )-lX
-n-:;l -n-a
-n-a
•
•
18
other possibilities,
~.~.,
a lower degree polynomial, were employed to
estimate the first two points but were rejected as no better than the
make-shift method outlined here.
observations
y*
Then, the vector of smoothed
can be computed by !* =
T
S
0
-1
S
-2
! , where
~
T
T
S
-:3
0
S
-4
S =
(nxn)
T
T
s
-6
T
s
-n-2
0
s
-n-2
o
s
-n-1
T
T
s T
-n
y( t ) •
y(t n )
2
J.
Similarly, by defining the (lX5) vectors
d T
-1
=
[0
1
2t
d T
-2
T
d.
=
[0
1
2t
=
[0
1
=
[0
=
[0
-~
d
-n-1
•
d T
-n
T
]
(X Tx )-lX T
<3
]
(X TX )-lX T
-3 -3
-a
2t.
3t.~ 2
]
(X. TX. f1 X.,
T i=3,4, ••• , n-2
1
2t
3t
1
2t
1
3t
1
3t
2
~
n-1
n
2
2
n-1
3t 2
n
2]
]
-:3
-a
-~ -~
-3
-~
Tx
)-lX T
-n-2 -n-2
-n-2
(X Tx )-lX T
-n-a -n-.
-n-a
(X
•
19
a matrix
D can be constructed so that
estimates of
.9L1
dt t=t.
y* = D Y is a vector of
derived from the 5-point moving-arc cubic
l
smoothing.
In an obvious manner, matrices
Sand
D can be constructed
for 5-point moving averages using linear-hyperbolic functions or
any power-polynomial function.
Further, it is clearly not necessary
to use smoothing functions of the same degree for all
The problem of choosing the
,
f(t.)
l
the
g(t.)
l
t.l
,
and
.
K
to
minimize equation (3.1) under the specification of 5-point movingarc smoothing and subsequent derivative estimation results in
minimization of
(3.4)
where matrices
Sand
D were defined above.
From equation
(3.4),
y(t ) (including
i
t , t , ••• , t ) and for a
n
1
:a
it is evident that, for a given set of values of
specific knowledge of the time-sequence
given moving-arc smoothing polynomial,
of
u:a
in
(3.4) involves choosing
However, since the
y(t. )
l
!.~.,
a cubic, the minimization
K to minimize the last term,
are assumed to contain an additive
random error, it is possible to attempt to reduce the effect of this
random error by applying the smoothing matrix
involving
K Y in
S to the terms
(3.4) so that the expression to be minimized is
(3.6)
It will be shown in Chapter
4 that the minimization of (3.6) instead
•
20
of (3.5) often does result in an estimator of
variance and which is more subject to bias.
K which has greater
However, for completeness,
the discussion of Chapter 4 will involve equation (3.6) with
references to equation (3.5) being immediate with
S = I , the
identity matrix.
3.2.2
Parameter Estimation with Simultaneous Estimation of
The case of 5-point moving-arc cubic smoothing of
y(t.), i=l, ••• , n,
1.
coefficients
involves the estimation of
(a ., a ., a ., a.}
01.
11.
21.
n
values
sets of the
of equation (3.2).
21.
n
K
Similar to
the situation discussed in Section 3.2.1, minimization of the
Sobolev norm-type expression under the condition of simultaneous
estimation of the 4n parameters
representation of the
f(t.)
1.
a ..
J1.
and
K
requires the
by
a .
01.
a .
[1
t.1.
11.
a .
21.
a .
31.
However, in this case the vectors
a .
01.
a .
11.
a .
21.
a .
21
are not estimated by the usual regression techniques.
Instead, if
smoothing is not attempted for the first-two and last-two
y(t.) ,
1.
the specification of 5-point smoothing suggests the formulation
•
..
21
1 t
1 t
1
2
t2 t6
a
1
1
a
t2 t3
:a 2
0
a
a
a
a
1
yet )
a
13
a
a
yet )
03
2lii
yet )
3:il
fi
yet )
2
04
yet )
14
:3
24
yet )
034
6
---y(t n _)
a
o
a
a
1 t
a
n
yet
)
n-a
yet
).
n-:a
y(t _ )
n 1
yet n )
o,n-:a
1,n-:a
:a,n-:a
O3,n-:a
[5 (n-4 )Xl ]
[5(n-4)x(4n-4)]
of the form
!
~ =
y ,
where the symbol
-~
indicates approximation
in the sense of the minimum Sobolev norm-type expression.
It will be
necessary in Section 4.2 to compute the inverse of the matrix
! T!
+
!' T!',
of n-4 (4x4)
these
where
~,
~
denotes
matrices.
d~'
dt;
Since we assume
the task involves the inversion
t. ~ t.
1
J
for
i ~ j
inverses exist.
3.3
A Function of Segmented Cubic Polynomials
The case involving initial 5-point moving-arc smoothing and
•
•
derivative estimation (Section (3.2.1)) involves representation of
the point functions
f(t.)
1
and
g(t.) , of equation (3.1) by linear
1
•
22
combinations of the observations, where the coefficients are uniquely
determined by the time-pattern of the observations.
f( t. )
in Section 4.1 that, in the case where the
l
It will be shown
and
are
g(t.)
l
derived from segmented cubic polynomials, the minimization of
equation (3.1) involves the simultaneous estimation of the
the parameters which characterize the cubic segments.
K and
In the
following Section 3.3.1, we define spline functions and discuss
several criteria for spline functions from the literature as motivation of our own development of the segmented cubic polynomials as a
In Section 3.3.2, a complete description is provided
special case.
for the construction of the cubic segments.
The computing procedure
K and the parameters which
to simultaneously estimate the
characterize the cubic segments will be developed in Chapter
4.
...
3.3.1 Least-Squares Spline Functions
The following definition of spline function follows that of
Greville (1969).
Let
t , t , ••• , t
a
l
t , t , ••• , t
l
a
n
be a strictly increasing
A spline function of degree
sequence of real numbers.
knots
n
is a function
Set)
m with
satisfying the following
conditions:
(1)
In each interval (t., t. ), i == 0,1, ... , n,
l
l+l
t
=
_co
0
and
t
n+l
polynomial of degree
(2)
for
=
co ,
Set)
is given by some
l
m or less.
and its derivatives of order
Set)
where
1, 2,
... , m-l
are
lThe spline functions we construct will be of interest only
t s t ~ t •
l
n
•
23
(-=, =) ,~o~o,
continuous on
t. E (a,b),
is a function of the
....m-l
class
Given the
S(t)
~
n
o
points
(t., y(t.)), i = 1, 2, ..• , n, with each
l
l
Greville (1968) shows that for arbitrary
l
smoothest function
po in t s (~ . ~ . ,
k
of the class
f(t)
f(t.) = y(t.))
l
l
which t'fits" the
is a spline function of degree
t , t , .0., t
having abscissa values
C
k < n , the
l:3
n
2k-l
as knots, where the
n
"smoothest" function is defined as that function which minimizes
the integral
Sb [f(k)(t)]2
dt
0
a
It is noted that, in general, each of the adjoining polynomial arcs
..
must have equal ordinates and derivatives of orders
1, 2, ••• , m-l
at each knot.
Reinsch's (1967) scheme for smoothing by spline functions relaxes
the fitting requirement
and seeks a function
f(t.) = y(t.)
l
l
f(t)
which minimizes the integral
t
S n[f(2)(t)]2
t
f ( t)
among all functions
n
1::
i=l [
where
•
f(t)
S
dt
l
e
C2 (t , t )
l
such that
n
f(t.) - y(t.)]2
l
l
S
S
Q'.
l
is a given number and the
is a cubic spline with knots
Q'.
l
are weightso
t , 000, t
l
n
He shows that
and, hence, that the
•
24
polynomial {cubic) segments join at their endpoints and that
and
fh
are contlOnuous.
f
,
f
I
,
There f ore, th e a d
···
h
JOlnlng
segmen t save
equa1
ordinates and first derivatives at each knot.
DeBoor and Rice (1968a, 1968b) consider the case where, for the
given data pairs
(to, Y.),
1
1.
a polynomial spline function of degree
m is "least-square-fitted" on each of the
where the
~i
partition the interval
coincide with any of the observed
[a,b]
t-values.
k
intervals
(~
If
)
i' ~i+l '
but do not necessarily
The authors consider
both the case where the number and placement of the knots
~o
1.
are
specified and the case where the number of knots is fixed but their
placement is determined by attempts to seek a minimum (integral)
least-square
t
n
S
•
t
[f(t) - S(t)J2 dt
1.
which, due to the discrete nature of the data, is evaluated by a
numerical procedure,
~o~.,
trapezoidal sums.
An approximation function which is closely related to the
least-square splines described by deBoor and Rice (1968a, 1968b) is
the "smooth" continuous function composed of segments of cubic
polynomials which will be characterized by the following points:
1
(1)
The segments must join at the common knots.
(2)
The first derivatives (with respect to time) of any pair of
adjoining segments must be equal at their common knot.
•
lThis particular characterization is derived from a more general
formulation developed by D. C. Martin.
•
..
25
The whole function of such cubic segments can be fitted to
the points
(t., y(t.»
1.
using ordinary least-squares
1.
techniques.
(4)
The choice of the domain of each of the
k
segments is left
to the discretion of the user to encourage adaptation to the
vicissitudes of the experiments originating the data.
It is natural, in the use of this function to approximate data, that
the first derivative of the function be used to approximate the
derivative
~
dt
at the points of observation,
It will be necessary to note, for the development of Chapter
4,
that the construction method described in the next section provides
that the parameters uniquely representing a segmented cubic polynomial,
given a set of points and the sequence of knots, are, in fact, the
values of the function and the values of the first derivative of the
function at the specified knots.
by the letter
These parameters will be denoted
B .•
J
3.3.2 Construction of the Segmented Cubic Polynomials
The function
i = 1, 2, ••• , k,
f(t)
is composed of
k
cubic segments,
f. (t) ,
1.
and is constructed so that a pair of adjoining
segments have identical ordinate values and first derivative values
at the common knot or "join-point".
•
f
f. (t * )
if
t < t*
f. (t)
if
t E Ct.*
fk(t k* )
if
t
1.
f(t) =
That is,
0
1.
could be defined
0
1-1
> t*
k
, t.* ]
1.
•
26
*~ i=O, 1, ..• , k, t *0 ~ t 1 , t k* ~ tm)
(t.;
where
knots which partition the interval
[t ,t J
1
m
are the user-specified
for the case where the
m
observations
{(t., y(t.); i=l, 2, ... , m, t. E[t , t J, t.
< t.)
~
~
~
1
m
J-l
J
are given. 1
To guarantee uniqueness, we specify that
Consider, for the present development, the i
f. ( t) : ;: a.
~
Obviously,
~o
~
~l
..
[0,1 J •
~a
*J.
[to*
).-1 ,t.]
B,y the transformation
).3
t
T ::;::
the subset
~a
~a
).1
segment, the cubic
+ a. t + a. 't 2 + a. t 3
+ 2a. t + 3a. t 2
f. : ;: a.
th
2k + 2 ~ m •
-
t.*).-1
t.* - t *
).
i-l
of the domain of
f
is transformed onto
Therefore,
*~
f.(T)=b.
+ b.
~o
J.l
T + b. Ta + b. TIS , T E (O,lJ
).3
la
and
dfi
df*
i / dt
:
;
:
dT
dT
dt
=
a / *
*
(b.J.1 + 2b.J.a T + 3b.J.3 T) (t.). - t.),-1 ).
t.*
,
).-1
Letting
we define
B , j = 2i-l, ,2i,2i+l, 2i+2, as
j
follows:
B.
2J.-1
•
t <
*J.-1 ) = f~(T)1
J.
T=O
::;:: f. (t.
J.
:=
b.
1Forpurposes of this discussion
and for t > t *
k •
to*
J.o
f
could be left undefined for
•
27
df.
I
df.*
-- -2:.
dt
t=t.*
B.
21
=
1-1
dt
d// d,. IT=O
*1.
1',(")1
'1-'1
df.1
= dt
B2i+2
It=t~
1
= b.1.0
d1'.* dt
= d// d,.
I
T=1
= b.11.
+ b.
Is.1
+ b.
11
1a
+ b.
1G
=
In the more convenient matrix notation
B.
o
1
a1-1
0
o
o
o
b.
1
b.
11
=
B.
1
B.
:a 1+2
o
b.
1
o
o
o
o
s.
o
o
b.
..3
-28.
3
b.
2
s.1
-2
21.+1
1
1
1:01
b.
13
so
10
b.
1.1
=
12
1.3
Therefore, for the
Ct.*
1.-1
1.
1.
m
i
21.-1.
B.1
s.
1.
observations in the i
2
th
+:3
interVal
, t .*] ,
1.
1
,.
b.
1.0
1
b.
1).
•
B .
b.
12
1
T
m.
1.
y
m.).-1, +1
or
y
B .
m.
+1
1-1
21-1
"
B .
1-3"~ +2T~
J
J
(T .-2T~+T~)S.
J
J
J
1
3T~-2T~
J
J
21
(-"~+T~)8 .
J
J
1.
..
B .
»l+a
which is of the form
B i
a
t.B.
-1-1.
with the obvious notation,
y
-1
m.
1.-1
B i
a
B. =
-1.
Y
B.
a 1.+1
B.
21.+a
and
•
--
~.
-1.
denotes the (m.x4) matrix.
1
-fil.
=
1
Y
m.
1
+1
~.~.,
•
29
•
Now, the matrix
•
is constructed so that to approximate the vector
~
~
-1
B
0
(m x4)
1
B
Y
B
Y
2
~
-a
t B
Y
1
3
a
a
3
,..,
,..,
(m x4)
=
1
o
B
Y
m
ak+1iI
[mX(2k+2) ]
Remark 1.
If the above equation is treated as a least-squares fit,
then solving the above linear system of equations for the
•
1
yield
Of course, the predicted values of the observations are
The matrix
t'
~'
.!..~.
,
::::
(d~
d,.
1
Remark 2.
1
are easily computed.
= a.10
is
-2,.. +3"~ ]
(6,. .- 6,.~ )/ s .
1-4,. .+3"~
J
W;
-1
J
J
J
The original coefficients
f.(t)
~
/ dt)
d,. ij
the jth row of the submatrix
f( -6,. J. -6,.~)
/ s.
L
J
/I
Y
=
is, then,
-
•
would
B.
a ..
1J
of the i
+ a. t+ a. t lil + a. t
11
12
J
1
th
J
cubic
3
13
Since the relationship between the cubic
"
B
.
Y
•
30
1
coe£ficients before scaling,
a .. , and the coefficients after scaling,
1J
b .. , can be expressed by
1J
a.
10
1
*
-to Is.
1-1 1
t~ a/s.a
-to* 3/ S. 3
J.-1 J.
b.
J.o
a.
J.1
0
lis.1
a
-2t.* Is.
J.-1 J.
3t.*a / S. 3
J.-1 J.
b.
J.1
1-1
1
=
=
a.
1a
0
0
a.
J.3
0
0
where
S.
1
0
t.*J.. - t.*
J.-1
=
3
3t.*J.-1 Is.J.
lis.J. a
,
l/Si
C.b.
-J.-J.
b.
J.a
3
b.
J.3
and since
b.
J.o
1
0
0
0
B
b.
J.1
0
s.J.
0
0
B
b.
1a
-3
-2s.
J.
3
-soJ.
B
b.
13
2
-2
S.
B
ai-1
ai
D.
= -J.
=
S.
J.
B.
-J.
ai+1
J.
ai+a
then
I
I
a
a
I
C
10
,''
1
o
(4x4 )t,
I
I
=
I
1
B
0
:
Ca
J
---.-_ .. - - - --I
I
D
I
I
I
:(4x4)
:
'h. __ ..J
:
B
a
a
:
(4x4) :
----r---------l
D
I
:
I
3
:
:-------.,(4x4):
,-----
': Dk
o
o
: (4x4)
•
1
(4x4 ):
- - - - - - , - - - --I
11
I
D
(4kX4k;)
[4k.X(2k+2) ]
B
2k+2
•
31
3.4 Comparison of Estimators
Classically, a useful comparative measure of the suitability of
/I
an estimator has been the mean-squared error, which for estimator
of
/I
K is defined to be
/I
in which the quantity
estimators
e[(K-K)2J
K-e(K)
of
K
and which has the property
is defined to be the bias.
Of several
K, the one which has the smallest mean-squared
error if often said to be the "best".
Obviously, by definition, the variance and bias of the random
variable
/I
K are non-random parameters associated with the distribution
of the given estimator
/I
/I
K where the random variables of which
K is
a function have a particular, though, perhaps, unknown distribution.
..
/I
Frequently, the estimator
K cannot be expressed in a form readily
e( ~~)
amenable to application of analytic methods of determining
/I
var (K)
mation to
and
In such cases, it is often possible to devise an approxi/I
K from which approximations to
analytically determined.
/I
e(K)
/I
and
var(K)
The appropriateness of such approximations
is usually verified by simulation studies for values of
region of interest.
estimates of
/I
e(K)
In the event approximations to
and
for specific values of
can be
/I
var(K)
K.
/I
K within the
/I
K are not made,
are often found by simulation studies
In this context, the simulation studies
frequently involve the generation of artificial observations with
additive error from a known distribution.
For example, the generated
observations could be of the form
•
y(t.)
l
= ~,(t.)
l
+ E(t.)
l
•
32
where
€{t.) - h[O,v]
l
cov[€(t.),
€(t.)] =
l
J
and
°
v
with either
=
constant or
A
K
The mean,
"K from m sets of artificial observa-
of the
tions is used as an estimate of
value of
"K
e{K)
"
for error-free values,
"v=o
denoted by either
K
or
~.~.,
"Kexact
" = (K-K"v=o )
K-e(K)
for a particular
y(t.) = y(t.)
l
l
v.
The
will be
The total bias (biastotal) is
+
"
(K"v=o -e(K))
and, from a simulation study, can be approximated by
A
K- K =
...
Hereafter,
"v=o
K-K
"A
(K- "K
) + (K
- K) •
v=o
v=o
,which refers to the bias due-to-estimation
procedure, will be denoted by
biaSmethod' and
"K
V=O
A
-K
refers to the bias due to error in the observations as
which
bias
error
While the mean-squared error may represent a useful measure for
comparing several estimators, it fails to provide information concerning
the relative magnitudes of the bias and variance.
Since interest is
often centered on the bias of an estimator, it may be worthwhile to
consider the MSE and a quantity which relates the MSE and the bias.
Either of the ratios
or
•
*
~ =
·
a
b lastotal
MSE
•
...
33
are such quantities.
Trivially,
MSE
=
~a
II
1
var(K:)
*
~
~
It is obvious that
=
ll'a
_.::..~-
1 + ~a
is a measure of
standard deviation, while
which is attributable to
biastotal
in units of
~* is the fraction of the value of MSE
biastotal.
It is entirely possible that, of two estimators, the one with
the smaller mean-squared error might actually have the greater bias,
and, hence smaller variance, than the other estimator.
.
assuming the
II
K:
In fact,
values are distributed normally, it is possible that
the bias of the former estimator might be so large that the interval
e(~) ± 2Jvar(~)
would exclude the true
K: , or, equivalently, that
the ,interval bounded by
e(~)
2Jvar(~)-K
±
J~ar(~)
would exclude zero.
e(~)
However,
- K±
2Jvar(~) = _-_b_ia_s_t~o~t~a~l
±2
Jvar(~)
•
= -~
indicating that the value of
~
± 2
provides information concerning the
•
34
effect of the relative magnitudes of the
biaS
total
var(~)
and
in the construction of confidence intervals.
Obviously, there is interest in attempts to decrease the MSE and
~.
The value of MSE can be decreased by decreasing one of either
"
or
var(K)
by an amount greater than the other is
increased, ~.~., by accepting an increase in
(biastotal)S
as the
penalty for a greater decrease in the variance.
"
Since the biastotal (K)
is defined to be
" = biasme th0 d (K)
"
bias t o t a 1 (K)
"
+ bias error (K) ,
it would appear naively that decreasing either
IbiaSerror
"
(~) I
would result in decreasing
However, it will be shown in Chapter
the
biastotal
biastotal
and
bias
or
.
IbiaSmethodl
or
Ibias error I
On the other hand, in the estimators to
be discussed, an increase in
Ibiaserrorl
"
by an increase in variance (K).
•
(~) I
I
4 that, in some cases where
can be achieved when either
in later chapters •
Ibiastotal
( ~)
are of unlike signs, minimum
error
are judiciously increased.
Ibiasmethod
is generally accompanied
Individual cases will be discussed
•
35
4.
ESTIMATING
K IN ~
dt -- Ky
Consider the simple differential equation
*" =
in which the value of
observations,
and
£(t.)
K is unknown and is to be estimated from
The objective will be to minimize the
Sobolev norm-type expression
'lfl
=
I:
i=l
uy (t . ) - f (t . ) J2
J.
with respect to
K, where
Usually
g
f
and
.
n
J~ + w[y(t. )-g(t. )]2}
I: (Cy(t.~ )-f(t.)
J.
J.
J.
i=l
n
=
J.
f
and
+
~
are as defined in Chapter 3.
g
contain parameters which must be estimated so that
structure of the functions
4.1,
f
and
the estimation of
least-squares methods so that
y(t )
i
f
and
K by minimizing
and
f
g
~ = ~!
with
g
and
f
and
•
II
that this procedure yields an estimator,
to non-negligible bias.
II
Of special
li are
f =
S Y
D defined as moving-arc smoothing and
derivative estimation matrices as in Section 3.2.1.
"
is
are chosen by classical
by knowledge of the time sequence.
~
(4.2)
are completely specified for
interest is the case where the (nXl) vectors
and
K depends on the
g.
implemented after the parameters of
given set of
(4.2)
W[ I<Y ( t J.. ) -g ( t . ) Ja }
the theory associated with the estimation of
In Section
n
y(t.) = y(t ) + £(t.)
~
i
~
{y(t ), y(t ), ••• , y(t )} , where
1
a
n
is a random error.
J.
(4.1)
= Yo'
Ky(t), y(O)
K, of
K which is subject
For several true values of
K are investigated where the errors
£(t.)
J.
It will be shown
K, properties of
are distributed normally
•
36
with proportional standard deviation and with constant standard
deviation.
In Section 4.2, the estimation of
with the choice of the parameters of
minimization of (4.2).
K is implemented simultaneously
f
and
g
by means of the
Two particular forms of
f
and
g
are
investigated:
(i)
f
is constructed to be the point function resulting
from 5-point moving-arc cubic smoothing and
(ii)
f
f
,
::::: g •
is a continuous and differentiable function composed
of cubic segments and
f
,
::::: g •
Wherever MOnte Carlo simulation is conducted, the necessary normal
deviates are constructed by the built-in uniform random number
generator function RAND of the PL/C computer program compiler
(Blankinship, 1971).
Twelve such pseudorandom numbers were summed
and their theoretical mean was subtracted from the sum to yield the
pseudorandom normally distributed deviates.
All computer programs
were compiled and executed on either the IBM 360/75 or its successor,
the IBM 370/165, digital computers operated by the Triangle Universities
Computation Center, Research Triangle Bark, North Carolina.
4.1
Estimation of
K and
Smoothing Operations Conducted Separately
4.1.1
General Development
If
•
•
w is arbitrary in units of time-squared and if
f
and
g
are
the point functions derived from 5-point cubic or 5-point linear hyperbolic smoothing and derivative estimation, then, as described in
Section 3.2, the value of
K which minimizes
va
in equation (4.2)
•
37
•
is that value of
K which minimizes the last term of equation (4.2),
(4.3)
For the sake of generality, this discussion will consider the case
where
rather than
e~pression
(4.3)
is to be minimized.
Taking
setting
=0
and solving for
K, we have
Alternatively, if
~
=
K §l
£t = Ky
differential equation
the observed values are
y(t.)
is an approximation to the
where
= y(t.)
T
Y = (y(t 1 ) ••• y~tn )}
-
and if
+ E(t.) , then
1 1 1
or
DY
where
10
•
~
=
(~-K£)!.
=
K SY + DE-KSE
=
K
SY + 6
Considered as a regression model with errors
the above approximation admits a least-squares estimate of
that
K which minimizes
~,
K, namely
•
38
.
•
the same quantity as in
(4.4) .
One has that, upon substitution into equation
(4.5),
(4.6)
T T
.
T T
TT
Y.. §. De: + ! §. Q;y, + ! §. De:
+--------------T T
T T
T T
T T
¥. §.
~
+ Y:. §. Se: + ! §.
~
+ ! §. Se:
Since
is an estimate of
K under error-free observations,
!.~.,
! = Y:. ,
then
For the case in which !
•
(-1) bias
error
contains random error, i.e.,
!
rQ
is given by the expected value of the second and
•
•
39
third terms on the right-hand side of equation
(4.7),
namely
•
(4.8)
since
A
e(K) ,
To derive an analytic expression for
it would be
necessary to compute the expected value of expression
approach is complicated by the appearance of the
(4.8).
£(t.)
1.
This
in first-
and second-powers and in cross-products in the numerators and
denominators of the fractions involved.
However, the expression
1
-..,...-----------=-.
. . . .- - -.......-------.T T
T T
TT
l. §.
~
+
2~
§. S £ + ! §. S £
permits an approximation to
terms involving
to
•
(4.8)
£(t )
i
(-1) bias
error
to be formed in which the
appear only in the numerator.
is of the form
where, for practical purposes,
x
could be either
The approximation
•
40
•
(i)
1
l.TST
_ S
~
or
•
1
(ii)
yTST'2L
__
S
(iii)
T
i
Z~~
Since the expression being approximated is a factor in the complete
expression (4.8), the choice of either (i), (li) or
(iii) must be
investigated from two aspects:
(a)
The suitability of the approximation to the factor (4.9).
(b)
The suitability of the resulting approximation of
A
K,
as in equation {4.7).
This investigation involves computing the value of (i) and the means
and standard errors of the sample means of the quantities (4.9) and
A
(ii) and (iii) as well as those of the resulting estimates of
which were made using (4.9), (i), (ii) and (iii), where
S
=
K
I
and
D is the matrix defined in Section 3.2.1.for initial derivative
estimation by 5-point moving-arc cubics.
For each of
K = -0.1, -0.2,
-0.3, -0.4, -0.5, 200 sets of hypothetical observations
(y ( t ~. ); t.~ = 1, 2, ... , 15} were constructed, where
and
•
•
e(t i ) ~ h [0, ( ~~o y(ti))i] •
y(t.)
~
= 1000
-Kt.
..
The results shown in Table 4.1
indicate that approximation (ii) is satisfactory to approximate the
fraction {4.9) and that approximation (i) is not unreasonable •
1
e
.
Table 4.1
,
..
Simulation results for approximations to
1
·e
Entries in subcolumns 2 through 7 are means.
yTST SY
A
Var(K)
,where computed, are in parentheses.
1
approximation (i)
yTSTSY
approximation (ii)
approximation (iii)
---
fraction
10- 7
-0.1
2.3145
A
K
-0.1020
approximation
10- 7
2.3300
A
K
-0.1025
approximation
10- 7
2.3049
(0.0029)
-0.2
4.8707
-0.2076
8.2205
-0.2994
4.9305
-0.2092
4.8362
12.197
-0.4016
8.2222
-0.3033
8.1272
17.648
-0.4851
(0.0071)
-0.2075
-0.2993
12.255
-0.4070
11.994
-0.4010
4.8727
-0.4914
17.343
-0.4848
(0.0069)
-0.1020
-0.2076
(0.0048)
8.2212
-0.2996
(0.0056)
12.207
(0.0062)
17.183
A
K
(0.0029)
(0.0055)
(0.0064)
-0.5
2.3147
(0.0048)
(0.0056)
-0.4
-0.1019
approximat ion
10-7
(0.0029)
(0.0048)
-0.3
A
K
-0.4019
(0.0064)
17.650
-0.4855
(0.0071)
+:-
I-'
•
42
Using the approximation (ii),
of
1
the estimator for
II
II
K (equation
4.7)
can be approximated by
II
K• K
+ K
exact
exact
1
+--
(4.10)
T
z.9L
where
C
= STS
and
B
= ST n .
As in equation
third terms of the right-hand side of
of
(-1) bias
error
•
equ~tion
the second and
(4.10) are an estimate
Now the expected value of the approximation
(4.10) can be expressed in terms of moments of
orders.
(4.7),
E
of the first few
Although the tedious details of the deviation are reserved
for the appendix, the expected value of the approximation to
II
K
+
•
exact
T a
(l. Cy)
(4 t (t y.c .. )S var(E.)
i
j
J J1
1
+ t t (C .. B.. + C. . a + C.. C. .) va.r{ E.) var{ E. ) }
"
1
J
11 JJ
1J
J1 1J
1
J
II
K is
•
43
- (-l:--T)
s (2
C
¥.. .1.
+ 2
+
r: r:
i j
(I:
1
r: (r:
i
r:"
J
y.C
., )(I: y .B .. )
J J1 . J J1
J
var{ e:.)
1
y . C . . ) (r: y .B. .) var{ e: . )
J J1 j J 1J
1
j
(4.11)
(C .. B .. + C. .B. . + C .. B. .) var{ e: .) var ( e: . ) }
11 JJ
1J 1J
J1 1J
1
/I
K, where
An approximation of the variance of
by equation
J
/I
K is approximated
(4.10), would involve oomputations of e(!TceeTce)S which
arise from the form of
indicate that
TIT
;r §. Sy
..
e(~S). Since the results shown in Table 4.1
is not a poor ~pproximation of
1
/I
K could be approximated by
the estimator
t
/I • K
/I
K
+ /IK
exact
exact
which, when squared, would involve computation of
complicated procedure.
/I
Kia
::!:
, a less
Indeed,
/I
/I
2K s
2K
exact
exact
T
T
T
T
T
(2;r
f.!
+
!
Ce)
+
(I Be + ! .!?l + ! Be)
T
T
/Is
K
exact
l~
l~
+
•
e(!TQ!!TCe)
/Is
K
exact
(y"T~)S
T
T
S
(2y" .Q! + ! Ce:) -
/I
2K
exact
(IT~)S
T
T
(2y" Ce + ! c e )
•
44
where, as before,
and
Therefore,
e(~a)
_
approx
~a
"exact
2K a
_
exact
"exact
T
l.9X.
2K
+
1
"
J
[
K
+
I: B.. var ( E .) +
. 11
1
T
'l.9X.
exact
T
C
l. '::.fZ
:Ii
"exact
[~e~act
]
C
a [4 I: (I: c ..y.) a var( E • )
i
'l.'::.fZ
j
1J J
1
]
.. Ci .) var( E.) var ( e: . ) ]
[t t (C .. C .. + C.1J.C.1J. + CJ1J
1
J
'.
11 JJ
J
1
2K
T
('l. ~)
{ 2 I: (1: c .. y . )( I: B . y ) var( E. )
. . 1J J
ml m
1
1
J
m
a
(4.12)
(t. ci.y.)(t
J J
+ [2 t
.
J
1
"
m
B. y ) var(Ei)J}
1m m
2K
exact
(l.T ) a
9l
I: I: (C .. B .. + C.. B' + Cj.B .. ) var(E.) V\3.r(E.)
i j
1.1. JJ
1J 1 j
1 lJ
1
J
[t (t B .. y.)a var(E.) + 2 t (t Bj.y.)(t B.
.
.
J
1
+
t (I:
i
j
J1 J
1
"
1
J
1.
J
m
y ) vadE.)
1m m
B.. y.)a var (E. ) ]
1J J
1
t t (B .. B .. + B. .B. j + Bj . B. .) var ( E.) var( e: . )
"
11 JJ
1J 1
1 1J
1
J
1
•
J
1
•
II
e(K)
The ~pproximate value of
(4.11).
can be computed u~tng expression
II
Hence, the variance of
II
var(K)
K can be approximated by
II
..1
e(K a )
(4.12a)
approx
Of course, the presence of
non-norma~
errors would require the
re-writing of equation (4.12).
4.1.2
Numerical Example
II
K
In this section, some of the properties of the estimator
{equation 4.5) are investigated by means of constructing values
{y(t
to
l
),
y(t:a)' ••• , y(t
= i}
by the solution
y(t)
= 1000
e
~ = Ky , y(O) = 1000 , and generating sets of artificial data
= 1,
t.~
all cases
and
lS
ti
):
€
2, ••• , 15}
(t.)
,..., h( 0, v) , where
~
~
[P
y(t i
100
(i)
v
(ii)
v ;:: constant
(iii)
v
Sand
r
where
=0
)
t
0
y(t.)
~
v
= y(t.) +
~
e(t.)
•
~
In
is either
or
Qr
,
Dare (15X15) matrices constructed for 5-point moving-
arc cubic smoothing and derivative estimation, or
S;:: I
and
D is
as defined above.
To investigate
biasmethod'
~exact
is obtained by equation
(4.5), the expected value of the qpproximate
•
bias
error
by equation (4.10), and the approximate variance of
(4.12a).
is obtained
~ by equation
To investigate the validity of such approximations, a
Kt
•
46
Monte Carlo simulation study was conducted in which 200 sets of
artificial data were generated.
For a given value of
K, the
same pseudorandom deviates were used to construct the
E(t.) •
~
1\
The means and variances of the
K obtained by averaging over
the 200 sets are included in the accompanying tables as follows:
_[p y(t
v -
i ) ] ,
, p
100
= 10
v = 900
-S as
expected(appro~):
Table 4.2
expected(approx): Table 4.6
5-point
cubic
smoother
simulation:
Table 4.3
simulation:
- = -I
Table 4.7
expected(approx): Table 4.4
expected(approx): Table 4.8
simulation:
simulation:
S
Table 4.5
Table 4.9
These results indicate the following:
(1)
•
•
There is very little difference between the
1\
the
var(K)
from minimization of
and
(KSY - Dy)T (K§1 - DY)
Ii
K and between
(KY _ Dy)T (KY - DY)
-...
--
-
(2)
Biastotal' although obvious, is small compared to
(3)
For each estimator
tends to increase as
~ as a function of K,
IKI
Jvar(~)
IbiaSmethodl
increases, indicating increasingly
poor approximation of the exponential function by the cubic
function.
•
•
..
.;
;
Table 4.2
Estimates using approximations of
for proportional data error.
A
A
K
K
exact
e( K)
bias
m
p
and var(~) following initial 5-point moving-arc cubic smoothing
error
A sample size of 200 is ass~~ed.
bias
= 10.
bias
e
bias
var(K)
.Jvar(R)
Jvar(~)
A
t
MSE
S
s*
-0.1
-0.1000
-0.1023
0.0000
0.0023
0.0023
0.001738
.0.0)+17
0.0029
0.001743
0.0552
0.0030
-0.2
-0.1998
-0.2027
-0.0002
0.0027
0.005221
0.0723
0.0051
0.005228
0.0374
0.0014
-0.3
-0.2990
-0.3012
-0.0010
0.09 2 9
0.0022
0.0012
0.009943
0.0997
0.0071
0.009944
0.0120
0.0001
-0.4
-0.3966
-0.3975
-0.0034
0.0009
-0.0025
0.016197
0.1273
0.0090
0.016203
-0.0196
0.0004
-0.5
-0.4914
-0.4908
-0.0086
-0.0006
-0.0092
0.025650
0.1602
0.0113
0.025735
-0.0574
0.0033
Simulation results for the estimation of K following initial 5-point moving-arc cubic smoothing for
proportional data error. p = 10. n = 200.
Table 4.3
K
*.
A
K
exact
Ii
K
bias
m
bias
e
bias
t
•
var (K)
~var(~)
Jvar(R)
MSE
S
S*
-0.1
-0.1000
-0.1030
0.0000
0.0030
0.0030
0.001766
0.0420
0.0030
0.001775
0.0714
0.0051
-0.2
-0.1998
-0.2091
-0.0002
0.0093
0.0091
0.004695
0.0685
0.00)+8
0.004778
0.1328
0.0173
-0.3
-0.2990
-0.3011
-0.0010
0.0021
0.0011
0.006333
0.0796
0.0056
0.006334
0.02-38
0.0002
0.0034
0.008222
0.0907
0.0064
0.008234
0.0375
0.0014
-0.0131
0.010300
0.1015
0.0072
0.010472
-0.1291
0.0164
-0.4
-0.3966
-0.4034
-0.0034
0.0068
-0.5
-0.4914
-0.4869
-0.0086
-0.0045
+"
~
•
,
EstL~ates
A
A
K:
K:
using approximations of
A
bias
and var(K:)
error
5-point moving-arc cubics for proportional data error. p
Table 4.4
exact
e(K)
bias
m
bias
e
bias
following initial derivative
= 10.
A
t
..
vadK:)
e~timation
by
A sample size of 20C is assumed.
Jvar(~)
Jvar(~)
MSE
S
S*
-0.1
-0.1000
-0.1012
0.0000
0.0012
0.0012
0.001794
0.0424
0.0030
0.001795
0.0233
a.co08
-0.2
-0.1998
-0.2012
-0.0002
0.0014
0.0012
0.005237
0.0724
0.0051
0.005238
0.0166
0.0003
-0.3
-0.2990
-0.2995
-0.0010
0.0005
-0.0005
0.009998
0.1000
0.0071
0.009998
-0.0050
0.0000
-0.4
-0.3964
-0.3957
-0.0036
-0.0007
-0.0043
0.016274
0.1276
0.0090
0.016292
-0.0337
0.0011
-0.5
-0.4912
-0.4891
-0.0088
-0.0021
-0.0109
0.025812
0.1607
0.0114
0.025931
-0.0678
0.0046
Simulation results for the estimation of K following initial derivative estimation by 5-point
cubics for proportional data error. p = 10. n = 200.
Table 4.5
A
K
K
exact
-0.1
-0.1000
-;;:
K
-0.1020
bias
m
0.0000
bias
e
0.0020
bias
A
t
0.0020
vur(K)
0.001736
••
~oving-arc
~var(~)
Jv[).r(~)
0.0417
0.0029
0.004685
0.1117
0.0123
0.006268
-0.0076
0.0001
NeE
0.001740
S
s*
0.0 1;('.0
0.0023
-0.2
-0.1998
-0.2076
-0.0002
0.0078
0.0076
0.004627
0.0680
0.0048
-0.3
-0.2990
-0.2994
-0.0010
0.0004
-0.0006
0.006268
0.0792
0.0056
-0.4
-0.3964
-0.4016
-0.0036
0.0052
0.0016
0.008107
0.0900
0.0064
0.008no
0.0178
0.0003
-0.5
-0.4912
-0.4851
-0.0088
-0.0061
-0.0149
0.010097
0.1005
0.0071
0.010319
-0.1483
0.0215
.00-
co
•
•
~
~
Estimates using
h
h
K
exact
e(K)
approxin~tions
bias
m
of
h
bias
bias
e
bias
h
t
var (K)
)var(K)
Jvar
(~)
MSE
~*
S
-0.1
-0.1000
-0.0998
0.0000
-0.0002,
-0.0002
0.000224
0.0150
O.OOll
0.000224
-0.0134
0.0002
-0.2
-0.1998
-0.1991
-0.0002
-0.0007
-0.0009
0.000838
0.0289
0.0020
0.000838
-0.0311
0.0010
-0.3
-0.2990
-0.2971
-0.0010
-0.0019
-0.0029
0.002389
0.0489
0.0035
0.002398
-0.0593
0.0035
-0.4
-0.3966
-0.3929
-0.0034
-0.0037
-0.0071
0.005891
0.0768
0.0054
0.005941
-0.0925
0.0085
-0.5
-0.4914
-0. Lf849
-0.0086
-0.0065
-0.0151
0.013516
0.1163
0.0082
0.013744
-0.1299
0.0166
Simulation results for estimation of K follo\~ing initial 5-point moving-arc cubic smoothing and derivative
estimation for the case of constant error variance. v = 900. n = 200.
Table 4.7
K
••
and var( K) followbg initial 5-poir.t movJ.ng-arc cubic
error
smoothing and derivative esUmation for the case of constant data error variance. v = 900. A sample
size of 200 is assumed.
Table 4.6
K
Jl
-;;
h
K
exact
K
bias
m
0.0000
bias
e
-0.0001
bias
h
t
-0.0001
var(K)
0.000234
r-:T
"Vvad K )
,j'!ar(~)
0.0153
0.0011
0.00023 u
~lSE
..
';>
s*
-0.0065
0.0000
-0.1
-0.1000
-0.0999
-0.2
-0.1998
-0.2009
-0.0002
O.OOll
0.0009
0.000838
0.0289
0.0020
0. 0008 39
0.0311
0.0010
-0.3
-0.2990
-0.2973
-0.0010
-0.0017
-0.0027
0.001517
0.0389
0.0023
0.001524
-0.069~
0.0048
-0.4
-0.3966
-0.3948
-0.0034
-0.0018
-0.0052
0.002646
0.0514
0.0036
0.002673
-0.1011
0.0101
-0.5
-0.4914
-0.48 in
-0.0086
-0.0073
-0.0159
0.005312
0.0729
0.0052
0.005565
-0.2182
0.0454
0'
•
.
•
Estimates using approximations of
Table 4.8
error
moving-arc cubics for constant error variance.
A
K
bias
K
exact
e(~)
bias
m
bias
e
bias
and
v
•
A
var(K)
following initial
= 900.
deri\~tive
estimation by 5-point
A sample of 200 is assumed.
A
t
var(K)
)var(R)
Jvar(~)
~-1SE
"
':>
t::*
-0.1
-0.1000
-0.0997
0.0000
-0.0003
-0.0003
0.000217
0.0147
0.0010
0.000217
-0.0204
0.0004
-0.2
-0.1998
-0.1985
-0.0002
-0.0013
-0.0015
0.000850
0.0292
0.0021
0.000852
-0.0514
0.0026
-0.3
-0.2990
-0.2957
-0.0010
-0.0033
-0.0043
0.002382
0.0488
0.0035
0.002401
-0.0881
0.0077
-0.4
-0.3964
-0.3899
-0.0036
-0.0065
-0.0101
0.006048
0.0778
0.0055
0.006150
-0.1299
o.Cl66
-0.5
-0.4912
-0.4800
-0.0088
-0.0112
-0.0200
0.013500
0.1162
0.0082
o. O13~-·{iC
-0.1721
0.0283
Simulation results for estimation of K following initial derivative estimation by 5-point
for constant error variance. v = 900. n = 200.
Table 4.9
K
A
K
exact
A
K
bias
m
bias
e
bias
A
0
'.
~
var(K)
)var(R)
moving-~rc
Jvar(~)
MSE
S
cubics
s*
-0.1
-0.1000
-0.09'l7
0.0000
-0.0003
-0.0003
0.000232
0.0152
0.0011
0.000232
-0.0197
0.0004
-0.2
-0.1998
-0.2003
-0.0002
0.0005
0.0003
0.000826
0.0287
0.0020
0.000826
0.0104
0.0001
-0.3
-0.2990
-0.2959
-0.0010
-0.0031
-0.0041
0.001!,88
0.0386
0.0027
0.001504
-0.1063
0.0112
-0.4
-0.3964
-0.3919
-0.0036
-0.0045
-0.0081
0.002582
0.0508
0.0036
0.0026h8
-0.1594
0.0248
-0.5
-0.4912
-0.4791
-0.0088
-0.0121
-0.0209
0.005106
0.0715
0.0051
0.005543
-0.2925
0.0788
/
/-
'V1
0
•
51
4.2 Simultaneous Estimation of K and
'.
the Parameters of the Smoothing Function
i
(
4.2.1 General Theory
4.1, minimization of the Sobolev norm-type e4pression
In Section
(4.2» for a specified set of observations resulted in the
(equation
problem of minimizing
~ w [Ky(t.) _ g(t.)J2
~
~
i
K only, since the parameters associated with the
as a function of
function
il>
g
were determined by the choice of li1p;proximating polynomial
and by the time-sequence of observations.
smoothing function
•
function
g
~herefore,
coq,siq.ered,
f
and the
In this section, the
a~sociated
derivative-estimating
are assumed to be linear in their parameters,
for the type of polynomial functions,
and
g
ca,n be represented by
respective~y, where
~
and
~
f
II
discussion,
df
= g(t)
dt
-
so that
suggested by the notation.
~
~
f , to be
~13
and
.! I~
,
m ~ n , matrices and
are (mXp) ,
is the (pXl) vector of parameters.
~j'
In all cases in this
and
~I
'j/
are re 1a t e d , as
The Sobolev norm-type expression can
be written
(4.13)
where, as before,
elements,
1
n
and
Y is the vector of
K y(t.) , in our case evaluated with the observed values
~
of the variables.
•
yT = [y(t.) ..• y(t)]
Since
(4.13) can be expanded,
•
52
•
%t = Ky
so that, in the simple case
~
•
(4.14)
Taking
0(" )
oK
0(")
and
o!
and setting the
res~ting e~pressions
equal to zero yields
o( lfi)
oK
=
21<:wyTy - 2w~T_;r... IT!
(4.15)
..
and
(4.16)
..
These normal equations
(4.16) can be written
tTt + wtlTt l
.@"
_w!/T~
I
I
--
-
(pxp)
I
(pXl)
- - -----1-- --- .. - -
_wyT t ,
I
..
=
--
I
"
WyTy
I
I
I(
o
(lXl)
(4.17)
[(p+l)X(p+l) ]
.
•
which is of the form A y
=
~
Assuming A- 1
G
exists and letting
G
).l
I
1a
-,.--1 .... --
GIG
21
I
I
22
•
53
be the partitioned inverse of
~,
then
•
Since
~ =1
G iTy from (4.17), it fQllows that
1- -
It
K=
or, for
w
(4.18)
= 1(time-unit)2
,
(4.l9)
where
As in Section 4.1 (expressions (4,6) and (4.7)), this form of
permits the construction of an approximate expression for
Specifically, if I
It
K
=
bias
It
K
error
Z + ! ' then
(Z + ~)T~(Z + !)
= ------(Z + ~) T!(Z + .5)
•
•
(4.20)
•
As discussed in Section 3.4, since
then
K: ="K:
"
\=xact
+ "K:
exact
(4.21)
the last two terms of which are an estimate of
(-1)
. bias error •
Further, if it is assumed, as in Section 4.1, that
is an adequate approximation to
1
"K:
then
~
can be approximated by
exact
+ ~
exact
(4.22)
•
the second and third terms of which are an approximation to
Since (4.22) is linear in the E(t.) and in integer
~
error
, the expected value of the approximate
powers of the E(t.)
~
(-1) bias
•
55
(-1) bias error
(and, hence, of the
appro~imation
to
"K
by (4.22))
can be computed by equation (4.11), provided the distribution of
E
is known or assumed.
1
If
1
is an adequate approximation to
y"T!:l..
"K
tnen, by expression (4.12a), an approximation to the variance of
can be computed.
4.2.2
Numerical ExamRles
In this section, some of the properties
of
the estimator
~
(equation (4.19)) are investigated from sets of artificially constructed
observations
ti
= 1,
y(t.) •
~
2, ••• , 15
For
sever~l
values of
K , values of y(t.)
,
~
are qomputed blf
= 1000e
y(t )
i
-Kt
i
y(t ) = y(t ) + e(t )
i
i
i
e(t.)
~ h(O,v) , where either
~
and sets of observations
(i)
v
=
[p
y(t i )]
2
were generated for
or
100
(ii)
v = constant
(iii)
v
f 0
or
=0
Two pairs of functions
f
and
g;
are considered.
First,
f
is
the function composed of three cubic segments with knots at 0.5, 5.5,
10.5, and 15.5 time units.
•
In this case, the matrices
Section 4,2.1 are defined in Section 3.3 .
t
and
I'
of
•
56
The second function
f
to be considered is the function derived
from 5-point moving-arc cubic smoothing and derivative estimation as
described in Section 3.2.2.
In this case, however, the vector
! of
observations in equation (4.13) is the (55Xl) vector which appears
in equation (3.7), so that the matrices
E and
F of equation (4.19)
are of dimension (55X55).
By matching columns of
duplicate elements of the
y( t.)
y ,matrices
E*
and
which appear
1
~n
E and
F with the
the (55Xl) vector
F* of dimension (15X15) can be formed by
arithmetic addition of elements of
E and l
so that
,.
where
.
yT
is the vector of 15 observations
y(t ) .
i
It is important to note that, for a given value of
random deviates were generated to compute the
K, the same
e(t ) .
i
The tabular results are shown as follows:
v =
lp Y(
100
•
J·
ti )
, p
= 10
5-point
cubic
moving-arc
expected(approx): Table 4.10
3 cubic
expected(approx): Table
segments
simulation:
simulation:
Table 4.11
4.~2
Table 4.13
v
:;:
constant
= 900
expected(approx): Table 4.14
simulation:
Table 4.15
expected(approx): Table 4.16
simulation:
Table 4.17
•
,
,.
•
'.
A
Estimates using approximations of bias
and var(K) for simultaneous smoothing by 5-point moving-arc
error
cubics and estimativn of K for proportional error 'rariance. p = 10. A sample size of 200 is assQ~ed.
Table 4.10
A
A
A
K
exact
e(K)
bias
m
bias
-0.1
-0.1000
-0.0997
0.0000
-0.0003
-0.0003
-0.2
-0.1999
-0.1994
-0.0001
-0.0005
-0.0006
0.000099
0.000264
-0.3
-0.2995
-0.2986
-0.0005
-0.0009
-0.00:.4
0.000829
K
-0.4
-0.5
-0.3981
-0.4945
Table 4.11
K
exact
-0.1
-0.1000
-0.2
-0.1999
-0·3
-0.4
-0.5
-0.0019
-0.0055
-0.0013
-0.0020
bias
t
var (K)
-0.0032
0.002468
-0.0075
neg
Jvar(~)
Jvar(~)
0.0099
0.0162
0.0288
0.0007
a.OOll
0.0497
0.0020
0.0035
S
S*
0.000099
-0.0302
0.0009
0.000264
0.000831
-0.0369
-0.0486
0.0014
0.002478
-0.0644
0.001+1
liSE
Simulation results for simultaneous smoothing by 5-point moving-arc cubics and estimation of
proportional error variance. p = 10. n = 200.
A
K
-0.3968
-0.4925
e
;;
K
-0.0987
-0.2004
bias
m
0.0000
-0.0001
bias
e
bias
A
t
var(K)
-0.0013
0.0005
-0.0013
0.0('1)4
0.000090
0.000253
-0.2995
-0.2990
-0.0005
-0.0005
-0.0010
-0.3981
-0·3994
-0.0019
0.0013
-0.0006
0.000596
0.000811
-0.0108
0.001208
-0.4945
-0.4892
-0.0055
-0.0053
~var(~)
Jvar(~)
MSE
0.0024
K for
S
s*
0.0095
0.0159
0.0007
0.0011
0.000092
0.000253
-0.1370
0.0251
0.0184
0.0006
0~0244
0.0017
-0.0410
0.0017
0.0285
0.0020
0.000597
0.000811
-0.02ll
0.0004
0.0348
0.0025
0.001325
-0·3107
0.0880
V1
-..J
•
..
Estimates using approximations of
Table 4.12
and estimation of
A
A
A
bias
and var(K)
error
K for proportional error variance. p
e(K)
-0.1
-0.1000
-0.0997
-0.2
-0.1999
-0.1992
-0.3
-0.2991
-0.2978
-0.0054
-0.0121
bias
m
0.0000
bias
e
t
~var(~)
Jvar(I)
0.000230
0.0152
var(K)
NSE
S
s*
0.0011
0.000230
-0.0198
0.0004
-0.0001
-0.0007
-0.0008
0.000793
0.0282
0.0020
0.000794
-0.0284
0.0008
-0.0009
-0.0013
-0.0022
0.002151
0.0464
0.0035
0.002155
-0.0474
0.0022
0.005323
0.0730
0.0052
0.005352
-0.071+0
0.0054-
0.012774
0.1130
0.0080
0.012920
-0.1071
0.0113
-0.3968
-0.3946
-0.0032
-0.5
-0.4913
-0.4879
-0.0087
-0.0034
Simulation results for simultaneous smoothing by three cubic segments and estimation of
error variance. p = 10. n = 200.
Table 4.13
;:
A
K
exact
-0.1000
ass~med.
-0.0003
-0.4
-0.1
= 10. A sample size of 200 is
-0.0003
-0.0022
K
bias
for simultaneous smoothing by three cubic segments
A
K
exact
K
K
bias
-0.0988
0.0000
m
bias
e
-0.0012
bias
A
t
var(K)
Jvar{~)
Jvar(~)
K for proportional
MSE
-0.0012
0.000237
0.0154
0.0011
0.000238
S
s*
-0.0779
0.0060
-0.2
-0.1999
-0.2025
-0.0001
0.0026
0.0025
0.000757
0.0275
0.0019
0.000763
C.Oj09
0.0082
-0.3
-0.2991
-0.2974
-0.0009
-0.0017
-0.0026
0.001485
0.03,'35
0.OO~'7
0.001·'191
-0.0(,,75
0.0045
-0.4
-0.3968
-0.3982
-0.0032
0.0014
-0.0018
0.001874
0.0433
0.0031
0.001877
-0.0!}16
0.0017
-0.4913
-0.4841
-0.0087
-0.0072
-0.0159
0.002'130
0.0493
0.0035
0.002683
-0.3225
0.09 t12
-0.5
••
~
V1
C))
•
••
..
A
Estimates using approximations of . biaserror and ·...ar( K)
cubics and estimation of I': for constant error variance.
Table 4.14
A
A
I':
K
exact
e(l':)
-0.1
-0.1000
-0.0997
-0.2
-0.1999
-0.3
-0.2995
bias
m
bias
e
0.0000
-0.0003
-0.1981
-0.0001
-0.2934
-0.0005
bias
for simultaneous smoothing by 5-point moving-arc
v
A
var (K)
t
A sample size of 200 is assumed.
~var(~)
Jvar(I)
0.0020
0.0001
S
;*
0.000004
-0.1487
0.0216
~1SE
-0.0003
0.000004
-0.0018
-0.0019
neg
-0.0061
-0.0066
neg
0.0116
0.0008
0.000441
-1.5090
0.6949
0.0162
0.0011
0.001746
-2.3707
0.8490
-0.4
-0.3981
-0.3825
-0.0019
-0.0156
-0.0175
0.000134
-0.5
-0.4945
-0.4615
-0.0055
-0.0330
-0.0385
0.000264
Simulation results for simultaneous smoothing by 5-point moving-arc cubics and estimation of
error variance. v = 900. n = 200.
Table 4.15
-;:
A
K
= 900.
K
exact
K
bias
m
bias
e
bias
A
var(l':)
t
-_..
~var(K)
Jvar(~)
MSE
K
for constant
~
s*
_-->------~---
-0.1
-0.1000
-0.0992
0.0000
-0.0008
-0.0008
0.000019
0.0044
0.0003
0.000020
-0.1818
0.0325
-0.2
-0.1999
-0.1979
-0.0001
-0.0020
-0.0021
0.000108
0.01011
0.0007
0.COO1l2
-0.2019
0.0392
-0.3
-0.2995
-0.2933
-0.0005
-0.0062
-0.0067
0.000340
0.0184
0.0013
0.000385
-0.3641
0.1166
-0.4
-0.3981
-0.3837
-0.0019
-0.0144
-0.0163
0.000626
0.0250
0.0018
0.000892
-0.6520
0.2979
-0.5
-0.4945
-0.4589
-0.0055
-0.0356
-0.0411
0.001341
0.0366
0.0026
0.003030
-1.1222
0.5574
V1
\0
•
'.
I'
A
and var (K) for simultaneOlls smoothing by three cubic segments
Estimates using approximations of bias
error
and estimation of K for constant error variance. v = 900. A sample size of 2(:(; is assumed.
Table 4.16
K
A
A
K
exact
e(K)
bias
m
bias
e
bias
A
t
var (K)
~varC~)
JvarCI)
NSE
S
s*"
-0.1
-0.1000
-0.0997
0.0000
-0.0003
-0.0003
0.000029
0.0054
0.0004
0.000029
-0.0559
0.0031
-0.2
-0.1999
-0.1986
-0.0001
-0.0013
-0.0014
0.000077
0.0088
0.0006
0.000079
-0.1600
0.0250
0.0234
0.0017
0.000573
-0.204.8
0.0402
0.0471
0.0033
0.002367
-0.2568
0.0619
-0.3
-0.2991
-0.2952
-0.0009
-0.0039
-0.0048
-0.4
-0.3968
-0.3879
-0.0032
-0.0089
-0.0121
0.0005 4 9
0.002221
-0.5
-0.4913
-0.4737
-0.0087
-0.0176
-0.0263
neg
•
61
These results suggest the following:
(1)
The methods involving simultaneous smoothing, derivative
estimation and estimation of
with smaller
biastotal
K yield estimates of
K
and smaller variance than the
methods involving initial smoothing.
(2)
For the case of proportional variance, the method of
simultaneous 5-point moving-arc cubic smoothing and
estimation of
biaStotal
K yields estimates of
and smaller variance than the method involving
three cubic segments.
However, for the case of constant
variance, the method of cubic
slightly better estimates •
•
K with smaller
s~gments
appears to yield
•
62
•
5.
•
Two natural
(1)
GENERALIZATIONS OF THE SIMPLE CASE
of the work iq Chapter 4 are obvious:
e~tensions
application of the simple case
form~ation
to a single
differential equation of more than one term and (2)
the
~imple
application of
case formulation to a system of differential equations.
In Section 5.1, the estimation of
and
K
1
K
2
in the model equation
y(t) = K y(t) + K z(t)
1
Ii
is investigated where the function
f
in the Sobolev norm-type
expression (equation (3.1)) is either the ~oint function derived from
initial 5-point moving-arc polynomial smoothing, or the function
derived from
simultaneo~s
moving-arc polynomial smoothing and estima-
tion of the
K. , or the function derived from the simultaneous
~
fitting of cubic segments and estimation of the
g = f
K. , and where
l
l.
Section 5.2 includes a discussion of the estimation of the
elements of the
(2X2)
matrix
[~]
=
K in the sy~tem
[::
The formulation leading to
::]
[:]
is shown to closely parallel that of
the simple case.
Section 5.3 is devoted to a discussion of the estimators of
and
•
K
Iil
in the system
y = - K Y + K z
.
z
1,
a
K
1
•
63
•
Consider the single
•
K
Estimation of
5.1
and
1
in
K
a
differentia~
y =Ky+l<z
1
a
equation
dy(t) = I< yet) + K z(t)
dt
1
a
where
and
K
1
values
K
1
and
K
2
2
are constants to be estimated given observed
= y(t.)
l
y(t.)
l
= 1, .. .,
i
I<
n •
and
+ E(t.)
l
z(t.)
= z(t.) + '(t.) ,
l
l
l
As in the previous chapter,
mo~ing-arc
following 5-ppint
discqs~ion
of estimating
polynomial smoothing and
derivative estimation will be followed by discussion of simultaneous
smoothing and estimation of
...
an,d
I<
1
K
2
for the cases involving
cubic segments and 5-point moving-arc cubic smoothing •
5.1.1
Estimation of
Fol~owing
K
Smoothing and Derivative
Initial 5-point Polynomial
Estimat~on
For the differential equation (5.1) above, estimation of
and
I<
a
I<
1
by minimization of the Sobolev norm-type expression (3.1)
following initial 5-point polynomial moving-arc smoothing and
derivative estimation reduces to the problem of minimizing the
term
n
w t
•
(y(t ) - g(t ))2
i
h:l
of equation (3.1) or minimizing
T
w 5 5
-
-
= w(K )Sy
,-
1
th~
+ I< SZ - Dy)T (I< SY + I< SZ - DY)
2-
where, as defined in Section 3.2,
•
equivalent expression
-
1--;
S
2-
is either the smoothing matrix
associated with initial 5-point moving-arc polynomial smoothing or
S = I.
Expression (5.2) could have arisen independently of the
•
64
•
Sobolev norm-type expression if the differential equation (5.1) is
written in the approximation form
•
D~ = K S~ + K S z
-
1 -
= S
Cl
[~~J
[ KK:]
p
and, subsequently, in terms of the Qbserved
D Y
!
where
= ~ + ..§. , and
I(
1
and
K
~
[! ~J
~ = ! +
w ~*T~*
We find that, since
of
=S
[
! ' and
I(Kaj
we find
•
=0
,
Z, as
vJI.*.
•
is, in fact, expression (5.2), estimates
w !T! •
to minimize expression (5,2), we take
Settiqg
+
and
~* ~ ~ ! - (K 1 S E + K2~)
derived from minim~zing w
those derived from minimizing
I
e*To*
are identical to
•
65
or
II
.-
-1.
yT
K
1.
~T~ [y ~J
J:;:
II
ZT
K:
~
yTST SY
yTSTSZ
-1.
[:~
STD Y
- --
yTSTDy
---
-., -
---
ZTSTSY
~?£TSZ
---
=
---
ZTSTDy
STS is symmetric, it follows that ZTSTSY
---
Since
(5.4a)
:c
yTSTSZ
---
so
d
= det
rv~]
[l~TSY
=ZTSTSZyTSTSY
----
~T~TSZ
ZTSTSY
---
_ ZTSTSYZTSTSY
-----
Therefore,
II
ZTSTSZ
K
1.
:;::
II
K
---
1
d
_yTSTSZ
---
2
ZTSTSZyTsTDY _ ZTSTSyZTSTDY]
----
-----
=
[ _yTSTSZyTSTpy + yTSTSYZTSTDY
---....,-
However,
y(t.)
= y(t.) + E(t.)
1 1 1
leads to the
•
e~pression
1
d
-...-_--
and
Z(t.)
= z(t.) + '(t.) , which
1 1 1
•
66
"
K
1
T T
T
T
T
~ ~ .§.( zy - l!:.. ) .§. ~
a
=
,
+T T
T
. T
T
~ ~ £ (~ "1 y:!:. ) ~ ~
b
=
where, with
a =
a
+ b
exact
Q = ~T£
{~T~(zyT -yz T)Cy
_
"K: 1
} { (~+P T~
[(~+~)(~+.:) T
(~+.:)(:.+~) T ] ~(~+.:) }
_
{~T2.(~T_ YZT)~} {(~+.i)TQ. [(~+~)(~+!)T
_
(~+~)(~+~)T] ~(~+.:))
and
'"K:
A similar expression for
2
length of the expressions for the
The complexity and
is obvious.
quantiti~s
a
and
b
preclude
efficient investigation of analytical forms of the expected value of
the
bias
error
, as we did in Chapter 4, necessitating investigation
of bias qy Monte Carlo simulation with repeated estimation of
K:
2
1
and
from artificially constructed data.
5.1.2
Estimation of
Estimation of
1
K
1
K
and
K
and
yet)
•
K:
=
?
K
2
with Simultaneous Smoothing
in the differential equation
K yet) + K z(t)
1
2
by minimization of the Sobolev norm-type expression (equation (3.1))
where
f
is either the function composed of three cubic segments or
the function derived from simultaneous smoothing and derivative
•
67
estimation and concomitant estimation of the
K. , and where
1.
= f'
g
can be accomplished by setting the appropriate partial derivatives
equal to zero and solving the resulting normal equations for the
unknown parameters.
ITt + wt,T I ,
The normal equations in matrix form are
w!,T
X
,
wi 'TZ
-- - - - - - -
I
- - - - - - ----
B
a
,.
wyT I ,
wyT Z
wyTy
where
t
wyT Z
wzTt'
wZTZ
=
B
0
s
~
-
tTy
1
1\
I
I
/\
B
1
K-
0
:3
is the matrix associated with the particular function of.
segmented cubics or 5-point moving-arc cubic smoothing as defined in
Sections 3.2.2 and 3.3.
The partitioned form of the matrix in the
normal equations leads to a natural partitioning of the usual inverse,
"T
K
so that the vector
" ,
= (K
1
"K )
a
can be expressed as (for
time-unit-squared)
-1
"K1
=
"K2
=
•
where
l::]
F Y
w
=1
•
68
F ::': ~/(~Ti + ~/T~/)-l~T
-
-
--
-
r--
-
Paralleling the argument in Section 5.1.1,
"K1
-
"K
=
ZTE{ZyT _ YZT) F y
~T~(ZyT _ gT) ~ !
yTE(ZyT _ YZT) FY
=
=
~T~(ZyT _ YZ'r) E y
:3
anp., so
=~
where
for
a
and
C and
b
F
a
+~
b
lexact
are as in expre~sions (5.7) upon substituting
for
B.
A similar expression is obvious for
5.2 Generalization to a
Syste~
of
Diff~rential
~
"K:
:3
Equations
Consider the system of differential equations
~
dt -- K1 y(t) + K:3 z(t)
•
•
dz ::':
dt
K:
and assume that observations
3
y(t) +
K:
4
z(t)
(y(t.), z(t.):
1
1
i ::': 1, 2, ••. , n}
are
•
69
given where
y(t.)
= y(t.) + E(t.)
1 1 1
Z(t.)
= z(t.) + '(t.) .
1 1 1
and
It follows from (5.11) that
l
a
z
K
a
1
K
2
K
a
z
a
l
3
z
K
4
where
·T
Y
=
T=
~
aT
=
and similarly for
(~(t 1 ) ~(t 2 )
(y(t ) y(t )
1
2
a
(a
z
and
a )
z.
For the system (5.11), we define the Sobolev norm-type expression
.
U*2 = t [(y(t.) _ f (t.»2 + w(y(t.)
1
1
1
g (t.»2
1
1
1
+ (Z(t.) - f (t.»2 + w(Z(t.) - g (t.»2J
1
where
1
.
y(t.) = K y(t.) + K z(t.)
and
f (t.)
1
1
and
1
and
11121
If the values of
SY
2
f (t.)
2
1
SZ, respectively, where
~
.
Z(t.)
2
=
1
K y(t.) + K Z(t.) .
13141
are the elements of the vectors
is the matrix associated with
initial 5-point moving-arc polynomial smoothing, and if
g (t.)
2
•
•
1
are the derivative estimates
DY
and
g (t.)
1
1
and
DZ, respectively, where
D is the matrix associated with
S as defined in Section 3.2.1, then
minimization with respect to the
K.1
of
•
70
+
(z- - -SZ)T(Z- - -SZ)
can be considered as a problem
+ w(K Y + K Z - DZ)T(K Y + K Z - DZ)
3-
4-
-
3-
4-
-
minimization of
invo~ving
w(K Y + K Z - Dy)T(K Y t K Z - DY)
1-
since the terms of
2-
-
1-
2-
U*2
involving
-.
S are independent of the
K.
l
•
As in the univariate case, the details of constructing the estimators
II
K.l
will actually be derived from minimizing
w(K SY + K SZ - Dy)T(K SY + K SZ - DY)
1-
2-
with reapect to the
-
K.
l
1-
2--'
-
rather than minimization of expression
On the other hand, we can write equation
(5.12)
(5.14).
in the approximation
form
K
DY
S Y
S z
0
0
2
=
Dz
1
K
K
0
0
E.l
S z
3
K
4
or, in terms of
Y and
Z ,
K
S Y
DY
S Z
0
0
1
K
2
=
•
DZ
K
0
0
S y
S Z
3
K
4
+ 0
•
71
where
Sand
D are the smoothing and derivative estimation matrices,
respectively, associated with initial 5-point moving-arc polynomial
smoothing and where
K
Se:
0
S'
1
0
K
2
6 =
K
[::]
Se:
0
0
(3
S,
K
4
One can obtain least-squares-type estimates of the
with respect to the
K.
l
,
K.
l
by minimizing,
the expression
T
eTc
DY
SY
SZ
0
0
DZ
Q
0
SY
SZ
~J
=
. DY
DZ
-[
which is formally equivalent to expression (5.15) for
squared.
SZ
o
w = 1 time-unit-
Therefore, estimators derived from minimization of (5.15)
will be identical to those obtained from minimization of
eTc
be referred to as "least-squares" estimators in Section 5.3.
and
we have
.
•
I
g
=
and will
Letting
•
72
yT
ZT
aT
aT
aT
;T
~T
ZT
y
- g
z
Taking
oC§.T~)
oK
=
- 2
+ 2
yT aT
~T aT
;T ~T
~T ~T
yT aT
~T ;T
•
o<.~T.§.)
oK
gTg
aT ~T
~T ZT
-
setting
gT~
-
11
[~]
l:
-Z a
a y
= a , and solving for -K , one obtains
~J
11
K
•
73
-1
1
y1
0'1'
ZT OT
OT yT
OT ZT
"
K =
~~ [:
Z
0
0
Y
yT OT
ZT OT
~T yT
~]
OT ZT
gT~
[~]
(5.16)
which is similar to equation
.
y = Ky.
(4.5)
for the simple differential equation
However,
-1
yT OT
;T ~T
OT ~T
~T ;T
=
gTg
[~
Z 0
Y
0
:]
Z1' OT
_yT OT
1
~T ZT
~T _~T
d
so
Z'r
"K =
1
d
OT
_~T ~T
T ZT
0
ZT _~T
gTg
[~ -:
o
~
0]
-I
yT OT
;T OT
OT yT
OT ZT
gT~
[~J
.
(5.17)
•
•
74
After multiplication,
~
:::
1
11
K
:::;
2
11
K
4
-1d
_ yTSTSZyTSTDY]
· [yTSTSYZTSTDY
----- -----
-1
:::
-1
11
K
•
d
:::
3
J
-1
-
~T§?SYZT.§.TDY
-ZTSTSZyTSTDZ _ ZTSTSYZTSTDZ
__ -----
L- -
d
d
[~?QTSZyT.£TDY
J
• [yTSTSYZTSTDZ - yTSTSZyTSTDZJ
----------
where
Alternatively, the approximation to the system (5.11) can be written
(sy)T]
[ (SZ) T
or, in terms of the observed
!
and
For purposes of this development,
•
(DY
i DZ)
(SY
! SZ)
~
~,
more convenient form is
•
[ ::' . :"4]
+ 6*
•
75
where
6* -- (De:
K
K
K
K
l
g)
~) - (Se:
:3
4
:il
In the preceding discussion, the estimates of the
K.
as elements of
l
a vector were derived from minimization of the sum of the squares of
elements of the error vector, viz.,
manner, estimates of the
K.
T
6 6
= t 6~.
l
In an analogous
in the matrix formulation above can be
l
derived by considering minimization of the euclidean (or Frobenius)
matrix norm
t
where
i,j
(6 *.. ) -- 6* = (DY:DZ)
-,lJ
-
_.-
(SY ! SZ)
K
K
K
:a
I<:
l
3
4
However,
6* = [DY - (SYK
-l
-- [
+ SZK )
2
DZ
oj(-
(SYK
-:3
6*
6
-l
+ SZK ) ]
-
4
]
--cl
so
t 6..*2
lJ
* 6* + 6*T 6*
= 6
--l
-:L.
--cl
-.a
which is formally equivalent to (5.15) for
.
•
Therefore, the estimators
"1<:.
l
w= 1
time-unit-squared.
derived from minimization of (5.20) are
identical to those derived from minimization of the Sobolev norm-type
expression (5.15).
•
76
•
.
In summary, the estimators
"K1
"K2
"K
and
3
"K
in the system
4
(5.11) are identical for each of the following formulations in the case
of initial 5-point moving-arc polynomial smoothing and derivative
estimation:
Consideration of each equation of the system separately
(1)
as in Section 5.1(2)
Minimization of the Sobolev norm-type expression (5.13).
(3 )
Consideration of the system in vector fo rmula t io n (5.12).
(4)
Consideration of the system in matrix formulation (5.19).
For the case involving simultaneous smoothing, derivative
..
estimation and concomitant estimation of the
functions
f
1
by the vectors
, f
2
,gl'
and
ts, ~ , !'~,
at the
K. , the values of the
1
t.
1
can be represented
and !'y, respectively, where the
matrix ! , whose elements are functions of the
t. , and
1
~
and
y , which are vectors whose elements are parameters characterizing
the smoothing function, are defined as in Sections 3.2.2 and 3.3.
For 5-point moving-arc cubic smoothing, the
~
and
yare the
coefficients of cubic polynomials; for cubic-segment smoothing, the
~
and
yare the values of the function and the first derivative
at the chosen knots.
.
•
Equation (5.13) is, then, of the form
+ (~ -~)
T
.
T·
(~-~) + w(~ - !'y) (~- !'y)
•
77
Taking
o(tf )
oK.1
and
..
letting
w
=1
time-unit-
squared, and setting the partial derivatives equal to zero yields
normal equations of the form
~T~ +
~ ,T~
a
- -- - - _yT~,
.
_ZT~,
aT
aT
/
_~
a
~T~ + ~ ,T~
-
-
-
<
,Ty
_~
,Tz
a
_~/Ty
a
_~
,T
"
~Ty
y"
~TZ
~
z
a
a
aT
yTy
yTZ
a
a
"K1
aT
yTZ
ZTZ
a
a
"K2
a
a
a
y'I;y
yTZ
"K
a
a
a
yTZ
ZTZ
"K
a
-
-
.
I
-
_yT~,
_ZT~,
=
(3
4
a
Clearly, from the blocked sYmmetry of the submatrices of the partitioned
matrix, the estimates of the
K.
from this formulation are identical
1
to those derived from consideration of the two differential equations
separately, as in Section 5.1.2.
5.3 Generalization to a System of Differential Equations in
Which Some of the Coefficients Are Related or Assume Known Values
..
As a particular case of the general system of differential
equations (5.11), consider the system
•
y = -K Y + K z
1
z
=
2
y(a)
K z , z(a)
:;3
=a
= z0
•
78
in the general form
y
•
y
z
0
~1
0
~.,
=
z
where
For
l:l
n
z(t.)
l
-
"'1 -
-
I<:
0
~
1
.,
=
y
0
I<:
observed values
= z(t.)
l
(5.21)
~3
.,
z:
S4
~3 = 0,
~ 4 =-1<: .,
([y(t.), Z(t.)J:
l
+ '(t.),
i
l
= 1,
= y(t.)
l
y(t.)
l
l
2, •.• , n}
+ E(t.),
l
and for initial 5-point
moving-arc polynomial smoothing and derivative approximation represented
by matrices
S
and
~
, equation (5.21) can be written in the
approximation form
Dy
S Y
S Z
0
0
13
~.,
=
Dz
1
+ 6
(5. 23)
~3
0
S Z
S Y
0
S4
where
D E
S E
~~
0
S'
0
0
S E
S'
6 =
0
In this section, it will be shown that estimates of
I<:
•
.,
=
13
.,
1<:1 ~-el
and
by the least-squares method as defined in Section 5.2 after
adjustment for the linear restrictions (5.22) are identical to
•
79
estimates of
K
and
1.
K
from application of the same least-squares
2
methods to
Q
!] !
[-.e.
[D Z
--
0
.e.
~]
[K
-8 Z
l
]
+ 0*
(5.24)
K
--
~
where
Goldberger
(1964) proves that, if £ is the unrestricted least-
squares estimator of
in the general linear model, of which our
~
b * , the restricted least-
equation (5.23) is a special case, then
squares estimator of
is
~,
(5.25)
where, in our case,
x
and where
r = R b
=
8 Z
o
o
8 Y
defines the exact restriction on the
K. .
1.
In
the case under consideration, the restrictions (equations (5.22))
r = ~~
have the form
I'l
•
[:1
=
[: :]
0
1
I'2
1
0
1'3
1'4
•
80
It is shown in the appendix that the second term in equation
(5.25) is a 4Xl vector, the first two elements of which are
and
1
d d
1
d
d
1
_ yTCZZTCY
= 2yTCYZTCZ
-----
2
yTCYZTCZ _ yTCZyTCZ
= ----
From equation (5.18), the unrestricted estimators
b
1
= [_~TCZyTBY
band
1
2
bare
2
+ ~TCYZTBYJ. 1
d
2
(5.27)
b
2
= [_rTCYZTBY
+ rTCZyTBYJ. 1
d
2
The sum of equations (5.26) and (5.27) yield the restricted estimators
•
•
81
b
*1
-- (-2~?CZyTBY + .~?CZZ TBy
1
1:TCZZTBZ ) •
d
1
(5.28)
b
(_~?CyyTBY + rTCYZTBY _ rTCYZTBZ)
*2 =
1
d
On the other hand, the least-squares estimators
equation (5.24) can be computed by minimizing
and
verified that, writing
B
=
1
'I(" 1
and
6*T 6*.
'K" 2
in
It is easily
T
S D,
'I"( 1
1
=
d
'K" 2
The estimators of equation (5.29)
are identical to those derived by restricted least squares, equation
(5.28) •
The estimation of the
Ie
1
and
of the Sobolev norm-type expression
K
2
in this case by minimization
(5.13), which for simultaneous
smoothing, derivative estimation, and concomitant estimation of the
K.
1
,
is
(_I(
1
•
Y + K Z _ !/~)
2
+ (~ _
! y)T
involves normal equations of the form (for
w
(~ _ ! y)
=1
time-unit-squared)
•
82
~T~ +
I
~ ,T~,
I
a
I
_~
~,Ty
,T
~Ty
"
Z
~
~,TZ
Y.."
I
I
~T~ +
a
- -
-
-
- -
-
-
-
-
-
-
-
-
-
-
~ ,T~,
-
-
yT~,
aT
_ZT~,
ZT IR ,
-
-
I
I
I
- 1- -
I
I
I
I
I
I
a
-
-
-
-
-
-
-
-
-
- -
~TZ
=
yTy
_yT Z
"K1
a
_yTZ
T
2Z Z
"K2
a
The solution of these normal equations yields
2Z TEZ
"K1
y'I'EZ
yT FY
1
=
"K
yTEZ
2
yTEY
_ZT FY + ZTFZ
d
1
d
=
where
Except for the definition of the matrices
E and
~,
these estimators
are identical in form to those derived from initial smoothing in the
preceding discussion (equation (5.2~)) •
•
•
83
6.
BIAS REDUCTION
6.1 General Considerations
From the discussion in Chapter
simple differential equation
negligible
biastotal
reduce
. 1
t OlJa
bias
\biaSmethodl
biastotal
y(t)
4 on the estimation of K in the
=
Ky(t) , the presence of non-
in the estimates is obvious.
Attempts to
in this case naturally involve attempts to decrease
and/or
Ibias
error
I.
Although these two components of
have been treated separately computationally in previous
chapters, the fact that they are related is obvious from inspection
of the simulation sampling results of Sections 4.1.2 and 4.2.2.
is not surprising, then that reductions in
by changes in
bias
biasmethod
It
are accompanied
error
In general, three approaches to reduction of
considered in this chapter.
First, a function
f
biastotal
will be
(as in the Sobolev
nOIT·-type expression (3.1)) can be chosen which better approximates
the underlying data-generating function and its derivatives at the
points of observation.
For instance,
f
might be chosen to be the
simultaneous 5-point moving-arc cubic rather than the initial 5-point
cubic smoother.
This approach also includes such procedures as
eliminating the points of known or suspected poor fit or poor derivative approximation, such as eliminating the first two and/or the last
two points in the case of initial 5-point moving-arc polynomial
smoothing and derivative approximation.
•
For example, the omission of
the smoothing (or approximation in the case of error,-free values
y(t.)) and the derivative approximation of the first and second points
l
•
84
in the case of initial 5-point moving-arc cubic smoothing for
K := -0.5
(see the fifth row of Table 4.5) yields the following estimates of
points omitted
"Kexact
none
-0.4912
-0.0088
yet )
-0.5057
0.0057
y{t ), yet )
1
:a
-0.5017
o .oorr
1
bias
K:
method
The second approach involves the utilization of auxiliary
information.
For example, such information can appear in the form
of knowledge of a second differential equation which contains
coefficients in common with the first.
.
An example will appear
later in this chapter •
The third approach depends on a reduction in the error associated
with the observations.
situations where
It is to be noted, however, that, in the
biasmethod
and
bias
error
are of unlike signs
(as in initial 5-point moving-arc cubic smoothing for proportional
error variances, results of which appear in Tables 4.4 and 4.5), a
minimum
Remark 1.
biastotal
can occur for nonzero
Although the estimator of
p.
K resulting from minimizing
the expression (4.4)
where
Sand
D are the smoothing and derivative estimation matrices
associated with initial 5·-point moving-are cubic smoothing, would
•
I
intuitively yield less bias than the estimator derived from mini.mizing
•
the approximation and simulation investigations of' 8eetio:l 4.1.2
suggest that this conjecture is not generall.y true.
Remark 2,
In the above example of decreasing
biaG me th
. 0d
by the
omission of the smoothing and derivative estimation at the first two
points, the assumed underlying function is
Although
y(l)
= 606.53
y(t):::: 1000 e- 0 . 5t .
y(2):::: 367.88
and
are used to smooth the
third point, the onission of these values from the formula for the
estimation of
the sample.
K results in disregarding part of the information in
In this particular case, the disregarded information
corresponds to an important part of the graph of the underlying
function, namely, that part where the values are greatest, where the
function is changing most rapidly, and where the probable errors are
greatest •
•
6.2 Examples of Reducing Biastotal
In this section, the estimation of
and
K
1
K
2
is considered
for the model differential equation
y(t)
=-
(6.1)
K y(t) + K z(t)
1
2
as a single equation from the underlying system of differential
equations
.
y( t) -- - K1"y(t) + K2 z(t)
I
(6.2)
z(t)
"
•
-
K z(t)
:a
For purposes of simulation, first, values of
computed for
t.l = 1,2, ... , 15
•
y( t. )
l
and
z(L)
l
from the solution of the system
are
(6.2)
•
.....
86
y(t) = z
0
['2 \,] [e-',
z(t) =
e
t -e
-K 2
t]
-Kt
where
z
o
= 1000,
K
~
Z
0
= 0.21
of Section 5.1, estimates of
2
,and
K
1
and
1
= 0.20
K
K
Using the
procedur~s
are obtained for the
:3
following smoothing and derivative approximation schemes:
(i)
Initial 5-point moving-arc cubic smoothing and
derivative approximation followed by estimation
of
(ii)
K
1
and
K
2
Initial 5-po~nt moving-arc linear-hyperbolic
smoothing and derivative approximation followed
by estimation of
(iii)
and
K
1
K:
2
Simultaneous smoothing and derivative approximation
using three cubic segments with knots at 0.5 , 5.5 ,
10.5 , and 15.5 time-units and concomitant estimation
of
(iv)
K
1
and
K
:3
Simultaneous smoothtng and derivative approximation
using 5-point moving-arc cubics and concomitant
estimation of the
K
1
and
K
2
To investigate the properties of
•
in the presence of proportional data error, Monte Carlo simulation
was conducted in which the artificial observations,
•
87
y( t. )
1
and
z(t.)
= z(t.) + ,(t.)
1 1 1
were constructed where
and
At the risk of the possible inclusion of peculiar (~.!., unlikely)
samples, the same sets of random deviates were used for both
and
....
p
= 10
and for all four estimation procedures.
that such duplicity assists in elucidating the
error and modifications in the estimators.
It is hoped
effe~ts
Table 6.1
results of estimation from error-free values,
=5
p
y(t. )
of increased
displ~ys
and
1
the
z(t.)
,
1
and from the sampling investigation.
From the
II
values (Table 6.1~, it is app~rent that the
Kexact
estimator associated with the initial 5-point moving-arc linearhyperbolic function is much poorer than those associated with the
other smoothing and derivative estimation functions.
the original values
y( t.)
1
and
z(t.)
1
Comparison of
and their approximations and
of the true derivative values and their approximations reveals that
approximations of
t. = 1, 2, and 3.
1
•
y(t )
i
and
z(t.)
1
are relatively poor for
The results of estimation of
smoothed values and derivative approximations for
are shown in Table
6.2.
K
and
1
t
i
=
K
2
using
4, 5, ••. , 15
These results indicate that the omission of
•
Table 6.1
..
t
l
Simulation results for estimation of
I<
1
and
I<
a
in
y=-l<y+l<z
a
1
~.
treated as a single equation from systerr: 6.2
Smoothing methods are initial 5-point moving-arc cubic = I-C, initial 5-point ~oving-arc linear-hyperbolic = I-L-H ,
simultaneous 5-point moving arc cubic = S-C, and simultaneous three cubic segments = S-3-C. K1 = 0.20 , K = 0.21
2
n = 25.
;;
K
~var(K)
.jvar(~)
-0.0598
0.000751
0.0274
0.0055
0.0021
-0.0652
0.001087
0.0330
-0.0612
0.0027
-0.0587
0.002862
0.2737
-0.0673
0.0036
-0.0637
0.004180
0.1986
0.1893
0.0014
0.0093
0.0107
0.2085
0.1979
0.0015
0.0106
0.0121
0.1986
0.1795
0.0014
0.0191
0.0205
bias
bias
p
K
K
I-L-H
5
K
0.2612
0.2598
-0.0612
0.0014
0.2773
0.2752
-0.0673
0.2612
0.2585
0.2773
1
K
2
10
K
1
K
2
I-C
5
K
1
I<
2
10
K
1
~*
0.004327
-2.1825
0.8265
0.0066
0.005338
-1. 9758
0.7964
0.0535
0.0107
0.006307
0.5463
0.0647
0.0129
0.008238
-1.0972
-0.98 4 5
0.000814
0.0285
0.0057
0.OQ0928
0.3754
0.1233
0.001104
0.0332
0.0066
0.001249
0.3645
0.1172
0.003217
0.0567
0.0113
0.00363'7
0.3616
0.0132
0.004873
0.3424
0.1155
0.1048
.~
0.4926
0.0226
0.004362
0.1988
0.1944
0.0002
0.0044
0.0046
0.000084
0.0092
0.0018
0.000106
0.5000
0.2005
0.2085
0.2038
0.0015
0.0047
0.0062
0.000136
0.0117
0.0023
0.000175
0.5299
0.2200
0.1988
0.185 4
0.0002
0.0134
0.0136
0.000310
0.0176
0.0035
0.OOO!~95
0.772'(
0.3734
0.2085
0.1950
0.0015
0.0135
0.0150
0.00049!~
0.0222
0.0044
0.000719
0.6757
0.3129
K
0.1991
0.1~'~1
0.0009
0.0050
0.000099
0.0100
0.0020
0.000134
I<
0.2089
0.2036
O.OOll
0.0053
0.0059
0.0064
0.000178
0.0133
0.0027
0.000219
0.5900
0.4812
0.2593
0.1870
0.1991
0.1831
O.OOO~
0.0160
0.0169
0.000381
0.0195
0.0039
0.000667
0.8667
0.4283
0.2089
0.1921
O.OOll
0.0168
0.0179
0.000680
0.0261
0.0052
0.001000
0.6858
0.3204
K
1
K
1
:3
1
2
10
or
VSE
0.0311
I<
5
t
0.0015
2
s-c
e
0.1874
K
10
m
~
0.2085
2
""
exact
bias
0.0660
II;
S-3-C
,---:::-
var(l<)
~
meth~d
K
1
K
2
co
(Xl
•
•
Table 6.2
Simulation results of estimating
and
K
1
K
2
in
y '"
KY + Kz
1
2
smoothing and derivative estimation for restricted sample.
n = 25
A
0.000273
0.0165
0.0033
0.000314
0.3879
0.1303
0.0113
0.000645
0.0253
0.0051
0.000772
0.4h66
0.1651.1
-0.0069
0.0035
0.001086
0.0330
0.0066
0.001099
0.10G}
O.Oll1
-0.01 2 5
0.0054
0.002625
0.0512
0.0102
0.002654
0.1055
0.0110
0.1921
0.1987
0.0179
-0.0066
K
0.1896
0.1965
0.0104
K
0.1921
0.2046
0.0179
1
2
= 0.21
~*
0.0064
2
0.20 , K
a
~
-0.0040
K
'"
MSE
0.0104
1
l
Jvar(~)
0.1936
IC
K
~var(K)
0.1896
5
bias
m
following initial 5-point linear-hyperbolic
t i '" 4, 5, "', 15 •
A
K
K
10
A
K
exact
p
bias
e
bias
t
'.
~
1
var(K)
0'
\.0
•
-
smoothing and derivative approximation at those points of poor fit
does result in reduced
biasmethod
and reduced
increased
bias
of opposite sign.
since the
biasmethod
biastotal
is reduced as the errQr rate,
error
and
biai'
error
I-
var(K)
but in
These results also show that,
are af unlike signs, the
p, increases.
As expected,
I-
the
var(K)
increases by approximately a factor of four when the error
variance is doubled.
in MSE, although
th~
This increase is also reflected in the increase
marked decrease in the values of
that the proportion of MSE which is due to
Although in a real
experiment~l
biastotal
~*
indicate
decreases.
situation it might be inappropriate
to consider the incomplete model consisting of one equation (6.1) when,
in fact, the complete model is known to be tne system (6.2), the
example presently under consideration (equations (6.1) and (6.2)) is
appropriate for demonstrating use of the method of
bias to tal
reduc-
tion incorporating a second differential equation with coefficients
in common with the original model equation.
example, estimation of
I(
1
and
K
2
For purposes of this
in the system (6.2) is conducted
utilizing the first procedure discussed in Section 5.3, which involves
the use of initial 5-point moving-arc cubic smoothing and initial
5-point moving-arc linear-hyperbolic smoothing.
estimation methods, and for each
For each of the
p; 5, 10 , the same sets of
random deviates are used to construct the errors
E(t.)
~
and
The sampling results for both methods are shown in Table 6.3.
'(t.) .
~
These
results indicate that, in the case of both smoothing functions, the
•
biasmethod
is reduced in comparison with the single equation
estimation (see Table 6.1), although the reduction is most drastic
•
..
•
Simulation results for estimation of
Table 6.3
K
1
and
K
2
•
\
in system 6.2 following initial smoothing and derivative estimation.
Smoothing methods are 5-point moving-arc linear-hyperbolic
K = 0.21 .
n = 25
= I-L-H
and 5-point moving-arc cubic
= I-C.
K1
= 0.20
,
2
method
p
K
I-L-H
5
K
1
K
A
K
exact
K
1
K
2
I-C
5
K
1
bias
m
bias
e
bias
A
t
var(K)
Jvar(~)
JV9.r(~)
~1SE
S
~*
0.2356
-0.0097
-0.0259
-0.0356
0.000731
0.0270
0.0054
0.001998
..;1.3185
0.6343
0.2184
0.2504
-0.0084
-0.0320
-0.0404
0.000904
0.0310
0.0060
0.002537
-1.3422
0.61+34
0.2097
0.2184
0.2431
-0.0097
-0.0334
-0.0431
0.002910
0.004768
-0.7996
0.3896
-0.0084
-0.0395
-0.0479
0.003695
0.0539
0.0608
0.0108
0.2579
0.0122
0.00~989
-0.7873
0.3831
0.1995
0.2036
0.0005
-0.0041
-0.0036
0.000502
0.0224
0.0045
O.C00515
-0.1607
0.0252
0.0049
0.000612
-0.1016
0.0102
0.2094
0.2125
0.0006
-0.0031
-0.0025
0.000605
0.0246
K
0.1995
0.2078
0.0005
-0.0083
-0.0078
0.002027
0.0450
0.0090
0.0020e8
-0.1733
0.0291
K
0.2094
0.2160
0.0006
-0.0066
-0.0060
0.002484
0.0498
0.0100
0.002520
-0.1205
0.01 113
K
2
10
K
0.2097
2
10
A
1
2
\D
f-'
•
92
in the linear-hyperbolic case.
bias
error
It is apparent, however, that the
increases in the linear-hyperbolic case while it decreases
for the cubic smqothing function.
Estimation of
K
and
1
~
~
6.2 by simultaneous
in the system
smoothing and concomitant estimation of the
K.
l
as discussed in
Section 5.3 results in estimates with insignificant
biastotal
and greatly reduced variance compared to single equation estimation.
The results of Monte
Ca~lo
simulation using the same sets of random
deviates as for the case involving initia+ smoothing (Table
appear in Table
•
6.4 .
6.3)
•
,
•
Table 6.4
Simulation results for simultaneous smoothing and estimation of
5-point moving-arc cubics
A
A
and
three cubic segments
1
and
= S-3-C.
A
K
1
K
2
in system 6.2
= 0.20
, K
:3
= 0.21.
~r-T
0.000062
-0.2078
0.0413
-0.1910
0.0350
0.000222
-0.0738
0.0054
0.0035
0.000309
-0.1322
0.0171
0.0055
0.0011
0.000030
0.0000
0.0000
0.0069
0.0014
0.000048
-0.1014
0.0102
0.0112
0.0022
0.000129
0.1875
0.0342
0.0138
0.0028
0.000191
0.0000
0.0000
K
S-3-C
5
K
0.1997
0.2016
0.0003
-0.0019
-0.0016
0.000059
0.0077
0.0015
K
0.20g5
0.2117
0.0005
-0.0022
-0.0017
0.000080
0.0089
0.0018
K
0.1997
0.2011
0.0003
-0.0014
-0.0011
0.000221
0.0149
0.0030
K
0.2095
0.2123
0.0005
-0.0028
-0.0023
0.000304
0.0174
K
0.1998
0.2000
0.0002
-0.0002
0.0000
0.000030
K
0.2097
0.2107
0.0003
-0.0010
-0.0007
·0.000047
K
0.1998
0.1979
0.0002
0.0019
0.0021
0.000124
K
0.2097
0.2100
0.0003
-0.0003
0.0000
0.000191
2
10
1
-
= 25.
0.000083
K
1
n
~*
p
K
Smoothing metr-ods are
~
method
exac
t
= S-C
K
•
bias
bias
bias
var(K)
'l/var(K)
·JvarCk)
m
t
e '
l>ll3E
:3
s-c
5
1
2
10
1
:3
\.D
'-"
•
7.
SUMMlI.RY AND OVERVIEW OF OPEN PROBLEMS
7.1
Summary
Consider a differential equation of the form
~-Ky+KZ
dt - 1
2
Given observed values
discrete times
t
1
,t
y(t.)
l
2
and
, ... , t
n
Z(t.)
l
of
y(t.)
l
1
z(t.)
l
at
,one is tempted to smooth the
data by fitting an approximating polynomial in
yT = [y( t ),
and
... ,
t
to the
y( t ) ]
n
ZT = [Z (t ), ... , Z(t ) ]
n
1
by a linear regression scheme, to estimate the derivatives
~
dt
t=t.
l
by computing the derivative of the approximating polynomial at the
t
i
' and to proceed to estimate the
K
i
by the usual regression
procedure applied to the model
y
•
by minimizing
T
~ ~
[ r*,:~*J
, where
Y*
+
and
Z
*
c
are the smoothed
This method often gives unsatisfactory estimates of the
•
the possibly poor approximation of the
y( t. )
l
Y and
K.
l
due to
from the observed
Z
•
95
values and the subsequent relatively poorer estimates of the
derivatives.
In this thesis, a method is proposed which is based on the
minimization of a quantity, the expression for which is related to
the discrete version of the Sobolev norm, namely
~)
where
f
and
g
are vectors, the elements of which represent the
values of the approximating and derivative estimating functions at
the
t.
l
and where
•
y
= [! zJ
[::J
Obviously, this procedure incorporates simultaneously the approximation of the
y(t.)
(or smoothing of the
l
y(t.)
l
in the case of data
with error), the estimation of the derivatives
~
dt
and the estimation of the
~
K
i
t=t.
l
By
suit~ble
f
and
, the naive regression-type approach is a special case of this
more general method.
g = f'
and
In each case discussed in this thesis, we have
f is a function which is linear in its parameters.
to the linearity of the parameters of
the coefficients
•
definition of
of equations .
f , and so, of
Due
g, and of
K. , the problem reduces to solving a linear system
l
•
96
In this thesis, three forms of differential equations are
considered in detail:
(1)
9X
dt
(2)
~==
dt
K Y
dy
dt
K Y + K z
1.
li1
== Ky
1.
+
K z
2
The general theory of estimating the coefficients is developed for
each and
the
K. •
1.
examples are includeq for certain representative values of
The approximation function
f
is either a 5-point moving-
arc cubic polynomial, a 5-point moving-arc linear-hyperbolic function,
or a recently-developed function composed of
joine~
cubic segments.
The numerical results, which include Monte Carlo simulation of
each of the above
d~fferential
equations with additive normally
distributed errOr, the standard deviation of which is either constant
or proportional to the
y(t.) , indicate that the minimization of
1.
tf yields substantially improved estimates of the
K.
1.
in terms
of less bias and less variance than the minimiz<:!-tion of the above
The estimation of the coefficients of a differential equation
by minimization of the Sobolev norm-type expression has several
..
advantages over other current methods.
know either the
•
ana~ytic
a numerical solution.
It is not necessary to
solution, if it exists, or the values of
Initial estimates of the parameters and
coefficients are not required, as they are in most iterative schemes.
•
97
Further, use of this method avoids, for the above described type, the
frequent convergence problems associated with iterative schemes.
The
procedure does not require a specific time-interval-sequence of
observations; in particular, it does not require equally-spaced
observations as do some transform methods.
On the other hand, disadvantages include the need for observations
on each variable in the differential equation at identical times.
Although the procedure is
comp~tationally
uncomplicated, the choice
of smoothing and derivative estimating functions is a potential
source of poor estimation,
•
7.2
Overview of Open Problems
This work suggests several general areas in which further
research is possible, including the followiqg:
(1)
Investigation of other choices of smoothing and derivative
estimation functions.
(2)
Extension of the basic method to different forms of the
model differential equation,
(3)
Investigation of minimizing a weighted Sobolev norm-type
expression.
Each of the above areas will be considered in more detail, although
their inter-relationships prevent
separa~e
and exclusive treatments.
In this thesis, we have considered two of the most common types
of smoothing and
11
•
d~rivative
polynomial function and (2)
estimation functiQns, (1)
the moving-arc
the continuous function defined over the
entire time-span of the data, an example of which is the function
composed of joined cubic segments.
Obviously, neither type of function
•
98
completely smoothed the observations and neither type provided exact
estimates of the derivatives.
This was hardly unexpected, since
polynomial approximations to exponential functions are not exact.
However, the use of approximating polynomials does involve the
assumption that the polynomial is a "suitable" approximating function
to the underlying data-generating function at least over the span
of fit.
Intuitively, the term "suitable" is associated with such
equally-undefined notions as "not distorting the data", whicn, for
example, can be interpreted as yieldiJ;1g values "very close" to those of
t:: underlying function.
Further, when derivative estimation is
accomplished by means of computing the first time-derivative of the
•
approximating function, the scope of meaning of "suitable" is
enlarged to include such features as preservation of monotonicity,
the presence of maxima (or minima), and the existence of asymptotes.
This type of derivative estimation has been iJ;1vestigated by many
authors (for example, se~ Carpahan, et ~., 1969), nearly all of
whom warn that small errors in the approximating function tend to
be magnified in differentiation.
One shoQld, then, strive for a balance on one hand of faithfulness to the underlying function and, on the other hand, of
reduction, by smoothing, of peculiarities in the observations which
are present due to error.
It is with these goals in mind that the
minimization of a Sobolev norm-type ex~ression (equation (3.1)) was
•
proposed •
Specifically, in the case of moving-arc polynomial approximations,
•
consideration
nomial~
mu~t
be given to the degree of the approximating poly-
the span of the arG, and the treatment of the first and last
•
99
groups of observations.
Most literature dealing with choosing the
appropriate degree for the approximating polynomial is based on
applications to physical experiments with many more observations than
we have assumed to be
availabl~
(~.g., see Savitzky and Golay,
from biological experimentation
1964).
Unfortunately, even the
number of points considered in one span is frequently larger than
the total number available from one biological experiment.
the case in the paper by Luers and Wenning
Such is
(1971) who discuss
derivative estimation by least-squares polynomial fitting of either
cubics or straight lines as a function of the
one span.
•
.
n~ber
of points in
Apparently, when relatively few observations are
available, the choice of both degree and number of points in a span
is as much art as science and is based mostly on the intuition and
experience of the experimenter.
In some biological applications,
knowledge of initial values suggests a particular smoothing and/or
derivative estimation function,
For the case
0::('
approximation by segmented polynomials, choices
involving the number and degree of segments and placement of the
knots must be made.
Some work has been done on choosing knots by
least-squares criteria, as in the papers by deBoor and Rice
and Hudson
(1968)
(1966), although in most cases it is still necessary for
the user to choose the knots, as Fuller
(1969) did, using the data
on which he reports.
Two potentially profitable directions for further research
include (1) .considerations regarding choice of norm and approximating
•
functions guided by graphical display and interactive computing
devices, which enable the user to make instantaneous decisions
•
100
(see LaFata and Rosen, 1970) and (2)
the approximating function, an
diversification of the form of
~xample
of which is the exponential
spline function developed by Spath (1969) to avoid undesirable
inflection pointE.
Another problem which would be worthy of more research involves
the parameter
w in the Sobolev norm-type expression (3.1).
the case
the
wh~re
K
i
For
are estimated following initial moving-arc
smoothing and derivative estimation, the value of
w is arbitrary.
However, in the cases involving simultaneous smoothing, derivative
estimation, and estimation of the
K ,
i
w has been set equal to
one time-unit-squared in t):lis thl;!sis without explanation.
This
parttcular choioe was made both on the basis of intuition and on
..
the basis of severa+ numerical studies which suggest that this
particular value is, in fact, reasonable.
The first such study
was one in which the single equation model (6.1) from system (6.2)
was investigated for
of the
K.
l
K
0,20 and
c
1
= 0.21 .
K
2
The estimation
was accomplished by simultaneous smoothing, derivative
estimation and estimation of the
t = 0.5 , 5 1 5
with knots at
Section 5.1.2.
K
i
using three cubic segments
10.5,
In this study,
and
15.5, as described in
w in the Sobolev norm-type
expression was allowed to vary,_ The results are shown in Table 7.1
and include
II
tfs = t.
K
lexact
l
.2'
lfd~ = w t [y(t.) - g(t.)J
.
l
•
l
l
2
~
and
suggest that the minimum value of
\f
tf
=
Cr(t.)
- f(t.)J2
l
l
u: + ~
These results
occurs for
time-unit-squared, at least for the values of
w equals one
w considered.
•
101
Table 7.1
Values of Sobolev norm-type expression as a function
of w • K = 0.20 and K = 0.21 •
1
,
2
0&
"K1
w
0.25
0.50
0.75
0.90
1.00
1.10
1.25
1.50
2.00
3.00
"K2
exact
0.1981
0.1984
0.1986
0.1987
0.1988
0.1988
0.1989
0.1989
0.1989
0.1991
1?
s
~
1?
2.586
2.936
3.197
3.319
3.390
3.454
3.539
3.655
3.655
4.042
13.226
12.243
11.816
11.666
11.591
11. 530
11.458
11.373
11.373
11.184
15.812
15.179
15.013
14.985
14.981
14.984
14.997
15.028
15.101
15.226
exact
0.2076
0.2081
0.2083
0.2084
0.2085
0.2085
0.2086
0.2086
0.2086
0.2088
In a second study, errorifree values of
y(t)
and
z(t)
were computed
for the same model equation as in the above example but for the values
of
K
1
and
K
and the time-sequence of observations associated with
a
the example included by Metzler (1969) in the descriptive manual for
the computer program NONLIN (see Table 7.2).
K
<3
Estimates of
K
l
and
were computed simultaneously with smoothing and derivative
t = 0.5 , 5.5
estimation by three cubic segments with knots at
20.5,
and
30.5
time-units.
The values of
"K
,
lexact
u: '
~,
values of
•
•
at
and
va ,
which are displayed in Table 7.3 for several
w, suggest that the minimum value of
w equals one
ti~e-unit-squared.
to investigate the influence of
"Ki
for data with errors.
tf
does occur
It would clearly be worthwhile
w on the quality of the estimators
•
102
Table 7.2
Values of
y(t.)
1
system (6.2) for
K = 0.1155.
2
....
•
1
from solution of
100 , K1 = 0.05775 ,
(From Metzler, 1969)
z
=
o
y( t.)
z(t.)
1.0
2.0
3.0
4.0
5.0
10.0
15.0
20.0
25.0
30.0
10.59
19.44
26.75
32.74
37.58
49.25
48.74
43.16
36.07
29.11
89.09
79.37
70.72
63.00
56.13
31.51
17.68
Table 7.3
0.25
0.5
0.75
0·9
1.0
1.1
1.25
1.5
2.0
3.0
z(t.)
t.
1
w
and
"K1
1
1
9·93
5.57
3.13
Values of Sobolev norm-type expression as a function
K = 0.05775 ,
K = 0.1155 .
of w
1
2
"K
exact
2exact
0.05661
0.05677
0.05687
0.05691
0.05693
0.05695
0.05697
0.05700
0.05703
0.05710
0.1147
0.1147
0 ..n47
0.1148
0.1148
0.1148
0.1148
0.1148
0.1147
0.1147
l?
~
if
0.007516
0.012562
0.016662
0.018746
0.020015
0.021204
0.022864
0.025374
0.0 29758
0.037194
0.081135
0.067005
0.060308
0.057770
0.056432
0.055298
0.053882
0.052049
0.049517
0.046485
0.088651
0.079567
0.076970
0.076516
0.076447
0.076502
0.076746
0.077423
0.079275
0.083679
s
•
'.
103
One generalization of the role of
w as a weighting coefficient
lead q to the following Sobolev norm-type expression
•
if
w
=
[(r _ !)T
.
T
(I - ~) ]
If we let the submatrices
w..
lJ
equal
w11
W
w21
W
V..
[~E-::-J
12
22
-1
where the
lJ
V ..
lJ
are
the submatrices of the variance-covariance matrix
v11
v =
V
lz
,
.'V
-----f-----
V
21:
ZZ
,
associated with the vector
then, for the case involving the model equation
observations
the
e(t.)
1
y(t.) = y(t.) +
e(t.) ,
y = Ky
i = 1, 2, ••• , n,
1 1 1
are independently distributed, and where
dimension (nXl), the matrix
W
11
where
Y is of
is diagonal with elements
(w11 )'i
= var(y(t.»)-l
1
1
•
Therefore,
w
: K-1W
11
•
with
,-
-- ---
K-1W
:
•
11
-.' - - - -
-
: K- 2 W
11 I
I
11
-
•
104
If we assume that
var(y(t.)) =
0
1
2
i = 1, 2, ••• , n , then
,
I
--
- - -,- - - - -1I I -2
I K
I
K
(~:
I
- ~)
I
The Sobolev norm-type expression minimized in this thesis is,
.
in the above form for the case
y
= Ky
,
I
t?
=
[ (~:
(~
!)T
~)TJ
I
I
I
I
(~
0
----1-----
- !)
-------
I
0
I
I
wI
(~
-
-
~)
I
which suggests that our expression, in the sense of this particular
case of variance-covariance weighting,. ignores correlations between
the
Y and the
Kr, and which suggests the role assumed by the
with dimensions identical to those of
K- 2
•
In fact, setting
w
w
equal to one time-unit-squared, as we have done, is tantamount to
designating
var(r)
=
I
and
var(Kr)
=
I.
Obviously, the development of estimators of
K by minimizing
a Sobolev norm-type expression with a weighting scheme is deserving
of future study.
One important aspect of such a study must be the
choosing of the weighting matrices.
Although the inverses of the
covariance matrices are used in the usual linear least-squares
techniques to insure minimum variance of the estimators for the
•
parameters of the model, it is not clear that use of the covariance
matrices in the Sobolev norm-type expression for the estimation of
parameters in differential systems will result in minimum variance
•
105
estimates of the
even exist.
K..
l
does not
Even if the inverses of the partitioned covariance matrices
are used as weighting matrices, the values of
must be known or estimated.
var{Y( t. ))
Y(t.)
l
1.
and of
Such estimation, if required, would
possibly be accomplished by an iterative scheme.
where the
V- 1
In fact, in the above example,
Further, for cases
are normally distributed and where the vector
Y
in the Sobolev norm-type expression is of dimension (mXl), where
m> n
as in simultaneous 5.. point moving-arc smoothing, derivative
estimation, and estimation of
K, the computation of the inverse
of the covariance matrices would probably require use of a form of
the generalized inverse.
A second major area of potential future interest involves the
extension of the basic method,
2:..~.,
minimization of a Sobolev
norm-type express.ion, to other forms of the model differential
equations.
It is our intention to include in this category both
the way the variables occur in the differential equation and the
form of the coefficients.
Perhaps, the simplest example is the
differential equation of the type
dy(t)
dt
K z(t)
which is often incorporated in models of the excretion phase of
pharmacokinetic studies (Cummings and Martin, 1964), where
..
•
is the amount of a drug metabolite in the body and
z(t)
y(t)
is the
amount of drug in the body.
Assuming that observed values of each
of the variables
are available at discrete times
y
and
z
t. ,
l .
then the Sobolev norm-type expression to be minimized is of the form
•
106
...
.,
from which the following normal equations are derived
~T~ +
- - - .-
W
- - - -
_wZT~
- I
t
t
1
- --I
I
I
"
-w ~/TZ
=
"
wZTZ
K
in the case
of the
0
yields
an expression which is similar to equations
"K
~Ty
~
----------
1
"K
Solving for
I
~/T~,
3l
= Ky
dt
•
(4.5) and (4.18) for
Obviously, the derivation of the estimate
K (equation 7.1) is straightforward; the investigation of the
properties of
"K
is made easier by the assumption of additive and
independent errors
e(t.)
l
and
'(t.)
l
associated with
y(t.)
l
and
z(t.)
, respectively.
l
A second example is the well-known system
proposed by Volterra in his study of fish populations (see Gael
(1971)).
al.
Depending on whether or not relationships exist between the
K. , estimates of the
K.
l
•
~
in Sections
the product
l
5.1 or 5.3
yTZ
= (x:
can be derived from the methods discussed
It is the presence of non-linear errors from
+
~)T(~ +
£)
which calls for further investigation.
•
.
107
A third related example which involves both the presence of
variables non-linearly and a non-linear relationship between the
K.1
is the differential equation
~
dt
= K Y
1
K
where
1
K =
2
, which Gause (1934) applied in his famous work on
K
2
yeast growth, where
y
is the size (in volume units) of a pure
culture of Saccharomyces cerevisiae.
From his observations, which
are shown in Table 7.4, Gause estimated
K ,the maximum population
2
size, by visual methods to be 13.0 volume-units and
"coefficient of geometr ic increase",
on
to be 0.21827.
t
the
K
1
by linear regre s sion of log [K ~ y]
2
Since equation (7.2) is linear in
K
K
K
and
1
~
, it is of interest to obtain estimates of
2
and hence, of
K
1
expression (3.1).
and
1
1
K
2
K ,by minimization of the Sobolev norm-type
2
K.
1
by one cubic segment with knots at
and 40 time-units, the estimates of
obtained.
and
K
For simultaneous smoothing, derivative estimation,
and estimation of the
•
and
1
K = 0.184
1
and
K
2
t = 6
13.28 are
This particular choice of knots, although it eliminates
consideration of the observations at
t = 48
and 53 time-units, is
justified by knowledge of the behavior of the cubic polynomial in
•
108
...
approximating a group of observations near the horizontal asymptote.
R by regressing
It might be worthwhile to compare the quality of
•
K2
log
[
-
y]
on
t
2
with our method.
Y
Table 7.4
Observations reported by Gause (1934) for the
growth of Saccharomyces cerevisiae.
time, hrs.
population size (volume units)
6
0.37
16
8.87
24
10.66
29
12.50
40
13.27
48
12.87
53
12.70
As a preliminary demonstration inspired by Gause's data, values
of
K
y(t)
=
2
were computed from equation (7.2) for
13.0 , y(6)
step-size two.
=
0.4
For the eight constructed values of
K
=
3
method.
•
and
and
K
and
1
above simultaneous procedure with
p
1
0.25 ,
by a fourth-order Runge-Kutta scheme with
in Table 7.5, estimates of
t = 5.5 , 27.0,
K
cubic segments with knots at
48.5 time-units.
5, estimates of
K
1
and
shown
were obtained by the,
2
t~o
y(t)
From artificially constructed
K
2
were computed by the same
The results are displayed in Table 7.6 and exhibit a strong
suggestion of significant bias in
"K1
,and, therefore, in
"K
2
Interest in models of the type (7.2) sho~ld probably initially
•
109
•
center on the effect of
square~
error terms, that is, on comparing the
influence of either squaring the smoothed
•
y( t. )
or smoothing the
l
y(t.):3 .
~
Table 7.5
Values of
of
y(t.)
computed by numerical integration
~
-YJ.
K
~:::Ky_:3~_
[
dt
1
K
K ::: 0.4 ,
1
K
:3
::: 13.0 .
:3
time
riJJ.
6
12
18
24
30
36
42
48
0.4
1.62
5.06
9.63
12.06
12.78
12.95
12·99
..
Table 7.6
Estimates of
K
and
1
K
:3
in
~ _ K
dt -
lY
Y]
K :3
[ ~K--
:3
Entries are means and standard deviations (in
parentheses) of samples of size n. K ::: 0.25
1
and K ::: 13.0 •
:3
II
p
n
o
3
20
5
25
•
•
K
1
0.2417
0.2358
(0.0143)
0.2208
(0.0271)
II
K
:3
13.087
13.160
(0.3350)
13.390
(0.5053)
•
110
•
The final example of the types of differential equations suggested
for future investigation is of the form
•
~=
dt
K
K
1
2
Y
which was suggested from empirical considerations in a study of water
vapor sorption in wood (Kelly and Hart, 1970).
in the estimation of the
K.
1
One possible approach
involves the use of transformations to
yield an equation of the form
log
~
dt
=
log K + K log Y
1
2
Another suggested area of further research is motivated by two
facts:
(1)
Sigmoid curves are often found in studies relating drug
dose and physiological response.
(2)
Certain differential equations are known to have solutions,
the graphs of which are sigmoid curves.
In many experimental situations, an estimate is sought either for the
dose which is lethal to 50 percent of the experimental subjects,
the so-called
LD
50
~.~.,
, or for the dose which induces a 50 percent-of-
Most current methods of
50
estimating these parameters rely on some scheme which purports to
maximum response,
~.~.,
the so-called
ED
linearize the dose-response relationship to permit the application of
simple linear regression techniques.
•
Besides often lacking theoretical
foundations, the basis of these methods frequently puts restrictions
on the inclusion of observations associated with 0 percent and 100
•
III
•
•
percent responses.
The estimation techniques advanced in this thesis
have potential value in deriving estimators of the dose-response
parameters, since, in many interesting cases, these parameters are
functions of the coefficients of the differential equations whose
solutions are sigmoid curves.
The final recommended area of further research is that which
would compare the methods of this thesis with those presently available
under the general title of non-linear regression procedures, which
involve minimizing the sum of squared deviations between either
analytical or numerical solutions to the differential equations and
the observed values •
•
•
.'
112
8.
LIST OF REFERENCES
Blankinship, S. (ed.) 1971. The User's Guide to PL/C. Document
Nwnber LSR-095-1, Triangle Universities Computation Center,
Research Triangle Park, North Carolina.
deBoor, C. and J. Rice. 1968. Least Square Cubic Spline Approximation
I: Fixed Knots. Report CSD TR 20, Computer Science Department,
Purdue University, Lafayette, Indiana.
deBoor, C. and J. Rice. 1968. Least Square Cubic Spline Approximation
II: Variable Knots. Report CSD TR 21, Computer Science Department, Purdue University, Lafayette, Indiana.
Cannon, J. R. and D. L. Filmer. 1967. The Determination of Unknown
Parameters in Analytic Systems of Ordinary Differential Equations.
SIAM Journal in Applied Mathematics 15:799-809.
,
Cannon, J. R. and D. L. Filmer. 1968. A Numerical Experiment on the
Determination of Unknown Parameters in an Analytic System of
Ordinary Differential Equations. Mathematical Biosciences
3:267-274.
Carnahan, B., H. A. Luther, and J. O. Wilkes. 1969. Applied Nwnerical
Methods. John Wiley and Sons, Inc., New York, New York.
Cleland, W. W. 1967. The Statistical Analysis of Enzyme Kinetic
Data, pp. 1-32. In F. F. Nord (ed.), Advances in Enzymology,
Volume 29.
Cook, D. A. and G. S. Taylor. 1971. The Use of the APL/360 System
in Pharmacology: A Computer Assisted Analysis of Efflux Data.
Computers and Biomedical Research 4:157-166.
Cummings, A. J. and B. K. Martin. 1964. Drug Elimination by
Simultaneous First-Order and Zero-Order Processes. Nature
202 :779-'780.
Evert, C. F. and M. J. Randall. '1970. Formulation and Computation
of Compartment Models. Journal of Pharmaceutical Sciences
59:403-409.
•
•
Fuller, W. A. 1969. Grafted Polynomials as Approximating Functions.
The Australian Journal of Agricultural Economics 13:35-46.
(Reprint Series Nwnber 255, Statistical Laboratory, Iowa State
University of Science and Technology, Ames, Iowa.)
Gardner, D. G. 1963. Resolution of Multi-Component Exponential
Decay Curves Using Fourier Transforms, pp. 195-203. In
Multi-Compartment Analysis of Tracer Experiments, Annals of
the New York Academy of Sciences, Volwne 108, Article 1.
•
113
Ga~se,
G. F. 1934. The Struggle for Existence.
Wilkins Company, Baltimore, Maryland.
The Williams and
Glass, H. I. and A. C. deGarreta. 1967. Quantitative Analysis of
Exponential Curve Fitting for Biological Applications. Physics
in Medicine a.nd Biology 12:379-3.38.
Goel, N. S., S. C. Maitra, and E.W. Montrol1. 1971. Nonlinear
Models of Interacting Populations. Academic Press, Inc.,
New York, New York.
Goldberger, A. S. 1964. Econometric Theory.
Inc., New York, New York.
John Wiley and Sons,
Greville, T. N. E. 1968. Data Fitting by Spline Functions.
Mathematics Research Center Technical Summary Report Number 893,
The University of Wisconsin, Madison, Wisconsin.
,
Greville, T. N. E. 1969. Introduction to Spline Functions, pp. 1-35.
In T. N. E. Greville (ed.), Theory and Applications of Spline
Functions. Academic Press, New York, New York.
Heinmets, F. 1970. Quantitative Cellular Biology: An Approach to
the Quantitative Analysis of Life Processes. Marcel Dekker,
New York, New York.
..
Hudson, D. J. 1966. Fitting Segmented Curves Whose Join Points
Have 1- Be Estimated. Journal of the American Statistical
Association 61:1097-1129.
,wksch, H. C. 1966. Random Errors of Derivatives Obtained from
Least Squares Approximations to Empirical Functions. SIAM
Review 8:47-56.
Kelly, M. W. and C. A. Hart. 1970. Water Vapor Sorption Rates by
Wood Cell Walls. Wood and Fiber 1:270-282.
LaFata, P. and J. B. Rosen. 1970. An Interactive Display for'
Approximation by Linear Programming. Communications of the
ACM 13:651-659.
Lewi, P. J., A. E. F. Chandler, D. G. VanRiel, P. J. Somers,
F. T. N. A11ewijn, J. G. H. Dony, and P. A. J. Janssen. 1970.
Computer Procedures for Evaluating the Results of Studies on
the Distribution and Metabolism of Drugs in the Body'.
Arzneimittel-Forschung 20:684-693.
Luers, J. K. and R. H. Wenning. 1971. Polynomial Smoothing Linear VS Cubic. Technometrics 13:589-600.
•
Martin, B. K. 1967. Drug Urinary Excretion Data: Some Aspects
Concerning the Interpretation. British Journal of Pharmacology
and Chemotherapy 29:181-193.
•
114
•
c. M. j969. A:j'isers ManuaJ, 1'0[' NONLIN. Technical Report
Numter 7292/69/7292/005, 'l'he Upjohn Company, Kalamazoo, Michigan.
MctzleL'~
Metzler, C. M., G. Matrone, and H. L. Lucas, J~. 1965. Estimation
of' 'l'ranspcrt Rates by Radioisotope Stlldies of J.'iro!J.~s t;eady-State
S;;rstems. Institute of Statistics Mimeo Sel-ies Number )+46,
North Carolina State University) Raleigh, North Ce.:c'olina.
Myhi11~
J. 196'7. Investigation of the Effect of Data. Error in the
Analysis of Biological I'race:..~ Data. Biophysical Joun'!.al
'7: 903-911.
l'1~rhil~tj J.
1969. Ir~vestigatiorl of the Effect of Data Error in the
Analysis of Biologica.l 'I'I-acer Data from a 'l'hree·-Compartment
System. Journal of Theoretical Bi.ology 23:218-231.
Parrish, J. D. and S. B. 8aila. 19'70. Interspecific Competition,
Predation and Species Di.versity. J'ournal of Theoretical
Biology 27:207-220.
,
Perl, W. 1960. A Method for Curve-Fitting by Exponential Functions.
IGter'national Journal of Applied Radiation and Isotopes
8:211-222.
Pizer, S. M., A. B. Ashare, A. B. Callahan, and G. L. Brow:J.el1. 1969.
Fourier Transform Analysis of Tracer Data, pp. 105-129. In
F. Heinmets (ed.), Concepts and Models in Biomathematics:-Simulation Techniques and Methods. Marcel Dekker, New York,
New York.
Reinsch, C. E. 1967. Smoothing by Spline Funct.ions,
Mathematik 10:177-183.
Numerische
Rescigno, A. arid G. Segre.1961. La Cinetica dei Far-mad e dei
Traccianti Radioattivi. Editore Boringhieri Societa per Azioni,
'l'or ino, via Brofferio 3.
English translation, 1966, P. Ariotti
(translator), Drug and Tracer Kinetics. Blaisdell Publishing
Company, Waltham, Massachusetts.
Rosenbrock, H. H. and C. Storey'. 1966. Corr~putat.ioGal Techniques for
Chemical Engic.eers. Pergamon Press, Oxford, England.
Savi.tzky, A. c..od M. J'. E. Golay. 1961.:.. Smoothing and Differentiation
of Data by Simplified. Least Squa:C'es Procedures. A.ile.ly'tical
Chemistry 36:1627-1639.
Spath, H. 1969.
4: 225-·233.
•
Exponential Splir:·e Interpolatior~.
Computing
[3wanr..., W. H. 1969. A Survey of' Non~Linear' Optimization 'l'echniques.
FEBS :L,etters, Volume 2 Supplement:S39~S55 .
1
•
115
•
Westlake, W. J. 19'71. Problems Associated with Analysis of Pharmacokinetic Models. Journal of Pharmaceutical Sciences 60:882-885.
Whittle, P. 1956. The Estimation of Age-Specific Infection Rates
from a Curve of Relative Infection. Biometrics 12:154-162.
t
•
•
116
9.
9.1
APPENDICES
Derivation of Equation (4.11)
The right-hand side of equation (4.10) is of the form
+
,
Since the
e(t i )
=
ei
are assumed to be independent, it follows tqat
the expected value of the second term is
[~e~ac~1 e
y"'CY
[~exac~ 1
::::
yT CY
= -
l~:;::t J
(~
T
1
T
Ce) +--e (~ Be)
yTCY
(~~
~
C.. var( e.) +_1_
i
"
1J
11
Now,
•
1
e
-· T C
Tc )2
(2--~
~- + ~ ~
c.. e.1 eJ.)
1J
1
e
+-yT cy
yTcy
(~~ B..
"
lJ
~ B..
.
1
11
1J
e. e .)
var( e.)
1
1
J
(9.2)
•
117
But
and
4Y TCe:Y TCE ::: 4[E e: :<3
+
j
T
T
e: Ce: E Ce:
- -
-
=
terms i.n
E.E
l
o
J
,
i
t-
j
,
E E E E E E E e:. C C.
km lJ
i j m k k m J l
0
E E e:.:<3 e: 2
i k l
k
C..
II
0
C + E E e:':<3
kk
i j
l
E. 2
C.. C..
J
lJ
lJ
+ E E Eo:<3 e: :<3 C . C' - 2 E E. 4 C.. 2
kl l kk
i k l
l
II
+
terms in odd powers of the
E.
l
from which it follows that
,
and
e
T
T
(~~ Ce:)
= E E
. k
l
+
c..
C
n kk
E E Ck,C' k
i
k
l
l
var ( E. ) var( E ) + E E
k
l
. .
l
•
First,
c.lJ.2
var ( E • ) var ( E. )
1
J
var(El.)var(E k )
The expected value of the fourth term of
similar manner.
J
(9.1)
is computed in a
•
118
Now,
e,
(2XTCEyTBe:) == 2
2
t (t
,
J
k
l
t (E
.
k
l
Ck'Yk)(I: B ..Y.) var(e: )
j
J
. lJ l
C . Y )( E B, ,Y ,) var( Ii: • )
kl k
, lJ J
l
J
and
e,
'I'
T
(~ Ce:e: Be:) ==
+ E E C. ,B. , var( e: . ) var( e: , )
,
l
+
,
.
J
Et
i k
lJ lJ
l
J
Ck,B'k var(e:l·)var(e: )
k
l
l
The sum of expressions (9.2) through (9.7), where each is
weighted as in expression (9.1), yields the desired re1ation~hip
(4.11) .
•
it follows that
'.
119
•
(~T~)-l
ZTCZ
_ZT CY
0
0
_yT CZ
yTCY
0
0
1
==
0
0
ZT CZ
0
0
_yT CZ
_ZTCY
yTCY
'Ci
2
Therefore,
~ (~~I'~) -1~ T
=
,
[~
1
:::
d2
0
1
1
0
~J
(~T~)-l
1
~ ~TCZ
_",TCY
_yT CZ
2yTCY
0
0
0
1
1
0
0
1
. 1d2
and
[~(~TK) -l RTJ
Tcy
-1
==
l2:f
ZT CY
1
(~T~)-l~T[~(~T!)-l~TJ
-1
1
- d
2
•
ZTCZ
0
T
-y CZ
0
yT CY
- CZ
-
_yT CZ
ZT CY
yT CY
ZT
_yTCZZTCY
==
l~l
--yTCYZ'l'CY
--
d
.d
2
1
lCy
r
lczJ.
ZTCY
ZT CZ
_yTCZZTCZ
---
yTCYZTCZ
--
1
d
1
d
2
d
1
120
»
However,
~TCZ;yTBX _ ~TCYZTBY
r - Rb
[~] [~
=
0
1
1
0
~]
yTCYZTBY _ yTCZyTBY
---
---
ZTCZyTBZ _ ZTCYZTBZ
---
---
1
~
I?CYZ T~ _ I?CZyTBZ
o
. -1d
=
o
so
_yTCZZTCY
---
=
1
d d
1
yTCYZTCY
---
_yTCZZTCZ
-~-
yTCYZTCZ
---
2
and the indicat~d product is exactly equation (5.26) .
•.
•
,
2