Reciprocal symmetry, unimodality and Khintchine`s theorem

Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
Proc. R. Soc. A (2010) 466, 2079–2096
doi:10.1098/rspa.2009.0482
Published online 15 February 2010
Reciprocal symmetry, unimodality and
Khintchine’s theorem
BY YOGENDRA P. CHAUBEY1 , GOVIND S. MUDHOLKAR2
AND
M. C. JONES3, *
1 Department
of Mathematics and Statistics, Concordia University, Montreal,
Quebec, Canada H3G 1M8
2 Department of Statistics and Biostatistics, University of Rochester,
Rochester, NY 14627, USA
3 Department of Mathematics and Statistics, The Open University,
Milton Keynes MK7 6AA, UK
The symmetric distributions on the real line and their multi-variate extensions
play a central role in statistical theory and many of its applications. Furthermore,
data in practice often consist of non-negative measurements. Reciprocally symmetric
distributions defined on the positive real line may be considered analogous to symmetric
distributions on the real line. Hence, it is useful to investigate reciprocal symmetry in
general, and Mudholkar and Wang’s notion of R-symmetry in particular. In this paper, we
shall explore a number of interesting results and interplays involving reciprocal symmetry,
unimodality and Khintchine’s theorem with particular emphasis on R-symmetry. They
bear on the important practical analogies between the Gaussian and inverse Gaussian
distributions.
Keywords: Cauchy–Schlömilch transformation; Gaussian–inverse Gaussian analogies;
Khintchine’s theorem; log-symmetry; R-symmetry; unimodal distributions
1. Introduction
R-symmetry and log-symmetry, which together fall under the rubric of reciprocal
symmetry, are to non-negative random variables what ordinary symmetry is to
arbitrary random variables. Inverse Gaussian (IG) distributions are as central to
the R-symmetric distributions as the lognormal distribution is to log-symmetric
distributions and the Gaussian (G) distribution is to the symmetric distributions.
Data from homogeneous, non-mixture, populations can usually be assumed to be
unimodal and the concept of unimodality of density was characterized in the 1930s
by Khintchine. These three concepts, reciprocal symmetry, G–IG analogies and
Khintchine’s theorem on unimodality, play basic roles in statistical modelling,
analysis and theory, and inter-relations between these three are the subject of
this paper. We begin with an extended introduction, looking at the definitions,
roles and practical importance of each of the three concepts in turn.
*Author for correspondence ([email protected]).
Received 15 September 2009
Accepted 14 January 2010
2079
This journal is © 2010 The Royal Society
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2080
Y. P. Chaubey et al.
(a) Reciprocal symmetry
Let Y be a random variable following an absolutely continuous distribution
function F with probability density function (pdf) f on the positive half-line R+ .
There are two natural and complementary concepts of reciprocal symmetry of f .
First, there is R-symmetry (Mudholkar & Wang 2007), which means that
q
,
(1.1)
f (qy) = f
y
for all y > 0 and some q > 0. Second, there is log-symmetry (Seshadri 1965),
so-called because it corresponds to ordinary symmetry of the distribution of
log Y . In density terms,
f (d/y)
,
(1.2)
yf (dy) =
y
for all y > 0 and some d > 0. While the importance of log-symmetry is obvious
through its transformational link with ordinary symmetry, the importance of
R-symmetry emerges as a driver of the extraordinary analogies between G and IG
distributions described in §1b to follow. Note that our use of the term ‘reciprocal
symmetry’ differs from its use in theoretical physics.
An initial comparison of these two notions of reciprocal symmetry was given
by Jones (2008). Before continuing to concentrate, in this paper, on differences
between R- and log-symmetry, it is worth recalling conditions under which R- and
log-symmetric distributions coincide; call these doubly symmetric distributions.
Trivially, this cannot happen if q = d. It is also easy to see that the lognormal
distribution corresponding to log Y ∼ N (m, s2 ) is both R-symmetric about q =
2
em−s and log-symmetric about d = em . Jones & Arnold (2008) showed that
absolutely continuous doubly symmetric distributions consist of a subset of those
distributions that have the same moments as the lognormal distribution. Write
k = d/q. Then, a doubly symmetric density f ∗ takes a piecewise form derived in
a closely related context by Pakes (1996), namely
f ∗ (qy) ∝
∞
k 2i(i−1) y 2i−1 j(k 4(i−1) y 2 )I (k −2i < y ≤ k 2−2i ),
(1.3)
i=−∞
together with the additional requirement that j(u) = j(1/(k 4 u)), 1/k 4 < u ≤ 1.
This class includes the lognormal, the Askey/Berg densities (e.g. Berg 1998) and
another example explicitly constructed by Jones and Arnold, but does not include
the famous class of Stieltjes densities (e.g. Heyde 1983).
(b) G–IG analogies
The basic paradigm for statistical theory and methods concerns questions
about location and variability. The IG distribution is a distribution for nonnegative random variables with two parameters, one being the expectation and
the other a measure of dispersion. This is a first respect in which it is similar to
the G distribution. The IG distribution has its origins in the studies of Brownian
motion by Schrödinger (1915) and by Smoluchowski (1915). However, its inverse
relation—in terms of Laplace transforms—to the G distribution was discovered
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
R-symmetry of probability distributions
2081
by Tweedie (1945) who gave it the current name. The G distribution has a
stronger following because of its physical explanations and numerous analytical
derivations summarized, for example, in Rao (1965). However, the IG distribution
has similar, numerous derivations and explanations, for example, by Schrödinger,
Smoluchowski, Tweedie, Wald, Huff and Halphen. These are summarized in
the historical survey chapter of Seshadri (1993). There is also a more recent
maximum entropy characterization of the IG distribution by Mudholkar &
Tian (2002), which is analogous to Shannon’s (1949) famous maximum entropy
characterization of the G distribution.
The G–IG analogies, in terms of both distributional and inferential properties,
are remarkable and manifold. Tweedie gave the earliest of these analogies,
which were highlighted in Folks & Chhikara (1976), a paper that made the
IG distribution better known and broadly used. The G–IG analogies have been
substantially extended and tabulated. For summaries, see tables 6 and 7 of
Mudholkar & Natarajan (2002), appendix B of Mudholkar & Wang (2007) and
table 2 of Mudholkar et al. (2009); the last two run to some 40 items. Mudholkar &
Natarajan (2002) also defined a concept of IG symmetry; this was mathematically
well defined but intuitively opaque. This led, in Mudholkar & Wang (2007), to
the alternative notion of R-symmetry, which offers some physical transparency.
Moreover, it then turns out that the R-symmetric distribution closely connected
to the IG distribution through which properties and consequences most readily
flow is the root reciprocal inverse Gaussian (RRIG) distribution. This is the
distribution of one over the square root of an IG random variate; its density
is given in equation (1.9).
The lognormal, the best known of the log-symmetric distributions for nonnegative data that is used in a variety of applications, has the advantages of
familiarity and simple transformation. On the other hand, the pivotal RRIG and
the related IG distributions, in view of the G–IG analogies, offer simple methods
for direct inference on population means without confounding by variability
aspects.
(c) Khintchine’s theorem
A simplified version of Khintchine’s formula for representing the distribution
function of a unimodal distribution with mode at 0 is given by
∞
Vz (x) dH (z),
(1.4)
F (x) =
−∞
where Vz (x) is the distribution function of the uniform distribution on (−z, z),
z > 0 and H is a distribution function. Khintchine (1938) identified the
distribution function H as
H (x) = F (x) − xf (x).
(1.5)
In the case that f (x) = F (x) exists and is continuous everywhere, the density
h(x) is given by
(1.6)
h(x) = −xf (x).
This leads to the following characterization of symmetric unimodal (with mode
at 0) distributions (Dharmadhikari & Joag-Dev 1988, p. 5).
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2082
Y. P. Chaubey et al.
Theorem 1.1 (Shepp 1962). If X is symmetric and unimodal (with mode at 0),
then F belongs to the closed convex hull of the set of all uniform distributions
on symmetric intervals (−a, a) with a > 0. Equivalently, if F is symmetric and
unimodal, then there exists a symmetric random variable Z such that F is
the distribution function of UZ , where U is uniform (0,1), independent of Z .
Alternatively, F is the distribution function of VZ , where Z is non-negative and
V is uniform on (−1, 1).
The basic characterization of unimodality owing to Khintchine (1938) received
enhanced attention when Gnedenko & Kolmogorov (1949) in their famous
monograph on the limit distributions of sums of independent random variables
used a false theorem owing to Lapin which claimed convolutions of real-valued
unimodal random variables to be unimodal. K. L. Chung, in his translation of
the monograph, highlighted the error via a counterexample and stated Wintner’s
(1938) result that convolutions of symmetric unimodal random variables are
symmetric unimodal (see also Feller 1971, p. 168).
The concept of unimodality of real-valued random variables, because of its
importance and ubiquity in statistical theory and applications, has received
considerable attention in the literature. It has been extensively studied, e.g. Laha
(1961), Medgyessy (1967), Sun (1967) and Wolfe (1971). It has also been variously
generalized, e.g. to alpha unimodality of Olshen & Savage (1970), generalized
unimodality of Ghosh (1974), multi-variate A-unimodality of Anderson (1955)
and S-unimodality of Das Gupta (1976). (For other technical details, see
Dharmadhikari & Joag-Dev (1988) and Bertin et al. (1997)).
Wintner’s (1938) result has been used by Birnbaum (1948) to analyse the
notion of peakedness. The result, when read to say that the area under a
symmetric unimodal curve over a symmetric interval decreases monotonically as
the interval is translated, yields monotonic power functions that are crucial for
setting sample sizes. Anderson’s (1955) theorem is a multi-variate generalization
of the earlier-mentioned reading of Wintner’s result (see also Sherman 1955).
It is used in Das Gupta et al. (1964), and later by many others, to study
power functions of multi-variate tests (e.g. Mudholkar 1965). Mudholkar (1966)
replaced the symmetry in Anderson’s (1955) result by a general group invariance
to obtain G-majorization and G-monotonicity related to majorization properties
(Marshall & Olkin 1979). Vitale (1990) offers a further generalization of
Mudholkar (1966). This result, among other applications, yields numerous
probability inequalities (Mudholkar 1969; Tong 1980; Dalal & Mudholkar 1988;
Dharmadhikari & Joag-Dev 1988) and is used for analysing multi-variate
peakedness (Mudholkar 1972).
(d) Further background and outline
Our starting point is the following:
Theorem 1.2 (Mudholkar & Wang 2007). Let f be the pdf of a unimodal random
variable that is R-symmetric about 1, then f belongs to the closed convex hull of
the set of all uniform distributions on R-symmetric intervals (1/a, a) with a > 1,
or equivalently, f is the pdf of UZ + (1 − U )/Z , where Z > 1 and U is uniform
on (0,1), independent of Z .
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
R-symmetry of probability distributions
2083
Theorem 1.2 for R-symmetric unimodal random variables may be seen in
parallel to Shepp’s representation of a symmetric unimodal random variable on
the whole real line. Mudholkar & Wang (2007) proved theorem 1.2 by establishing
the following Khintchine-type representation of the distribution function of an
R-symmetric random variable unimodal around 1,
∞
Wa (x) dQ(a),
(1.7)
F (x) =
1
where Wa (x) represents the distribution function of a random variable distributed
uniformly on (1/a, a), a > 1 and Q is a distribution function defined on the
interval [1, ∞). Mudholkar & Wang (2009) used theorem 1.2 to establish the
monotonicity of the power function of the test of significance for the IG mean.
In the remainder of this paper, we treat the canonical cases in which Y follows
distributions with q = 1 or d = 1. In this case, unimodal R-symmetric distributions
will have their modes at 1. The general cases can be reconstructed from these
by considering the distributions of qY and dY , respectively. That is, all results
remain valid if q = 1 or d = 1 provided the distributions are rescaled appropriately.
For example, in theorem 1.2, if f is R-symmetric about q, then it is the pdf of
VZ + (q − V )/Z , where V is uniform on (0, q).
Two particularly important R-symmetric distributions, mentioned earlier, that
will be considered from time to time in what follows are the lognormal distribution
which, when q = 1, has density
1
1
2
2
(1.8)
exp − (log x + m ) x > 0
fL (x) = √
2m
2pm
and the RRIG distribution (Mudholkar & Wang 2007) with density
l
1 2
2l
exp −
x−
x >0
fR (x) =
p
2
x
(1.9)
when q = 1; here, m, l > 0.
In §2, we present an alternative form of theorem 1.2, identify the distribution
Q for a given R-symmetric pdf f and study its consequences. Section 3 gives
a one-to-one correspondence between a symmetric family of distributions (on
R around 0) and an R-symmetric family (defined on R+ with mode at 1)
and explores further relations with log-symmetry and R-symmetry. Conclusions
follow in §4.
2. A Khintchine-type theorem and its consequences
(a) A Khintchine-type theorem for R-symmetric distributions
Theorem 2.1 is very similar to theorem 1.2, which is theorem 5.4 of Mudholkar &
Wang (2007). The only difference is that in theorem 1.2, Z has support (1, ∞)
while in theorem 2.1 Z has support (0, 1). The theorem provides a striking and
attractive property of R-symmetry.
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2084
Y. P. Chaubey et al.
Theorem 2.1. Let f be the pdf of a unimodal random variable that is
R-symmetric about 1, then f belongs to the convex hull of all symmetric densities
on R-symmetric intervals [a, 1/a] with 0 < a < 1 or, equivalently, f is the pdf of
X = U /Z + (1 − U )Z , where Z is defined on (0, 1) and distributed independently
of the uniform (0, 1) random variable U .
Similar to equation (1.7), the distribution function F (x) of an R-symmetric
unimodal random variable can be expressed as
1
(2.1)
F (x) = Wz (x)dQ(z),
0
where Q is a distribution function defined on (0, 1) and Wz (x) is the distribution
function of a uniform (z, 1/z), 0 < z < 1, random variable. The following theorem
addresses the 1–1 correspondence between the distributions of X and Z .
Theorem 2.2. Let X be a non-negative random variable that is R-symmetric
and unimodal about 1, with distribution function F and pdf f , then the random
variable Z is uniquely defined by its distribution function Q and pdf q (assuming
that f exists) given by
1
1
Q(z) = F (z) + 1 − F
+
− z f (z), 0 < z < 1
(2.2)
z
z
and
1
q(z) =
− z f (z),
z
0 < z < 1.
(2.3)
Proof. Consider some distribution Q on (0, 1). Then, we can use the
representation in theorem 2.1 to write
⎧ x
(x − z)
⎪
⎪
dQ(z),
if x < 1,
⎪
⎨ 0 (1/z) − z
F (x) =
1/x
⎪
⎪
(1/z) − x
⎪1 −
⎩
dQ(z), if x ≥ 1.
(1/z) − z
0
Also,
f (x) =
min(x,1/x)
0
1
dQ(z).
(1/z) − z
(2.4)
We find that the Radon–Nikodym derivative of Q with respect to f is 1/x − x.
Therefore, for 0 < x < 1,
x 1
Q(x) =
− z df (z).
z
0
Integrating by parts and noting that f (1/x) = f (x) and that xf (x) → 0 as x → ∞,
we get equation (2.2) for all continuity points x on (0, 1). At the remaining points,
we determine Q by right continuity, proving that F determines Q uniquely. In
particular, if f (z) = F (z) exists and is continuous almost everywhere, then Q
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2085
R-symmetry of probability distributions
has the density q given by equation (2.3). It can be directly checked that q
is a density for any unimodal R-symmetric f ; its non-negativity corresponds
to the unimodality of f and the demonstration of its unit integral uses the
R-symmetry of f .
Alternative approach. A derivation of theorem 2.1 and formula (2.2) that
parallels the development of Khintchine’s theorem by Jones (2002) provides
a useful geometric explanation of various aspects of unimodal R-symmetric
distributions. Let X ∼ f , f being the pdf of an R-symmetric unimodal random
variable with mode at 1, so that f (0) = f (∞) = 0, and define random variable
Y by
(X , Y ) ∼ fX ,Y (x, y) = I (0 ≤ y ≤ f (x)),
where I (E) is the indicator function of the event E. Furthermore, let
f (x) = f (x)I (0 < x < 1) and fr (x) = f (x)I (1 < x < ∞). Then, we have the
following lemma:
Lemma 2.3. (i) The conditional distribution of X given Y is given by
1
X |Y = y ∼ U f−1 (y), −1
f (y)
(2.5)
and (ii)
U
(2.6)
+ (1 − U )Z , where Z = f−1 (Y ).
Z
Proof. Note that owing to unimodality of f , it is increasing for 0 < x < 1 and
decreasing for 1 < x < ∞. Thus, fr and f are invertible and we have fr−1 (y) =
1/f−1 (y), 0 < y < f (1). From this, it is easy to see that equation (2.5) holds and
thus, unconditionally, that equation (2.6) holds.
X=
Lemma 2.4. For 0 < y < 1, the distribution function of Y is given by
1
1
−1
−1
FY (y) = F (f (y)) + y
− f (y) + 1 − F
.
f−1 (y)
f−1 (y)
(2.7)
Proof. For 0 < y < f (1), we have
y ∞
I (0 ≤ z ≤ f (x)) dx dz
FY (y) =
0 0
∞
=
min{y, f (x)} dx
0
=
f −1 (y)
0
f (x) dx +
1/f −1 (y)
f−1 (y)
y dx +
∞
1/f−1 (y)
Equation (2.7) now easily follows from the above equation.
≤ z) = P(f−1 (Y ) ≤ z) = FY (f (z)),
f (x) dx.
using lemma 2.4, equations
As Q(z) = P(Z
(2.2) and (2.3) follow.
The following examples provide the densities of Z corresponding to the familiar
R-symmetric lognormal and RRIG distributions.
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2086
Y. P. Chaubey et al.
12
µ=2
µ = 1.5
µ=1
µ = 0.5
µ = 0.1
µ = 0.01
10
qL(z)
8
6
4
2
0
0
0.2
0.4
0.6
0.8
1.0
z
Figure 1. Pdfs qL for m = 0.01, 0.1, 0.5, 1, 1.5, 2.
Example 2.5. For the lognormal distribution, fL (x) = (− log x)fL (x)/(mx) and
hence, for 0 < z < 1
1 1 1 − z 2 (− log z)fL (z)
qL (z) =
2
mz
1
1 1
2
2
2
=
(2.8)
1 − z (− log z) exp − (log z + m ) .
2m
2pm3 z 2
This pdf is plotted for different values of m in figure 1, which shows that the
density of Z is more concentrated near 0 for large values of m and near 1 for small
values of m.
Example 2.6. For the RRIG distribution, fR (x) = l 1 − x 4 fR (x)/x 3 and so for
0<z <1
1
qR (z) = l 4 (1 − z 2 )(1 − z 4 )fR (z)
z
2 1
2l3 1
l
z−
(1 − z 2 )(1 − z 4 ) exp −
.
(2.9)
=
p z4
2
z
It is interesting to note that if Z ∼ qR , then Y = l(Z − 1/Z )2 follows the c2
distribution on three degrees of freedom. Figure 2 gives plots of qR (z) for various
values of l which shows that Z is more concentrated near 0 for small values of l
and near 1 for large values of l.
(b) A link between R-symmetric and log-symmetric random variables
Theorems 1.2 and 2.1 are both analogues of Khintchine’s theorem for
R-symmetric random variables that differ only in the support of the random
variable Z ; in theorem 2.1, the support is (0, 1), whereas in theorem 1.2 it is
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2087
R-symmetry of probability distributions
qR(z)
15
l = 0.01
l = 0.1
l = 0.5
l=1
l = 1.5
l=2
10
5
0
0
0.2
0.4
0.6
0.8
1.0
z
Figure 2. Pdfs qR for l = 0.01, 0.1, 0.5, 1, 1.5, 2.
[1, ∞). However, these two can be combined to express a unimodal R-symmetric
Y in terms of a uniform (0, 1) random variable U and a log-symmetric random
variable Zm as at equation (2.10).
The link between the two versions of reciprocal symmetry relies on the fact
that if f is R-symmetric, then f (y) = −f (1/y)/y 2 , which leads to the following:
Theorem 2.7. An R-symmetric random variable Y can be expressed as
1
1
Y = U max Zm ,
+ (1 − U ) min Zm ,
(2.10)
Zm
Zm
in terms of a uniform (0,1) random variable U and an independent log-symmetric
random variable Zm , where Zm is Z with probability 1/2 and Z1 = 1/Z with
probability 1/2.
Proof. Now, while the density of Z in theorem 2.2 is given by equation (2.3),
which is on support 0 < z < 1, the density of Z1 ≡ 1/Z can readily be shown to
have the same form, just on a different support:
1
− z1 f (z1 ), 1 < z1 .
q1 (z1 ) =
z1
If we now define Zm as in the statement of the theorem, its density is
1 1
− zm f (zm ), 0 < zm ,
qm (zm ) =
2 zm
(2.11)
and Zm has the complementary property of log-symmetry! This can be easily
verified as 1/Zm has the density given by
1
1
1 1
1 1
− 2 =
− zm f − zm f (zm ) = qm (zm ).
2 zm
zm
zm
2 zm
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2088
Y. P. Chaubey et al.
Concretely, combining theorems 1.2 and 2.1, an R-symmetric random variable,
Y , can be written in terms of a uniform random variable, U , and the above
log-symmetric random variable, Zm , via equation (2.10).
Remark 2.8. If f (.) is smooth and well behaved in the sense that f (1) = 0,
f (1) < 0 and f (1) is finite, then the densities associated with Zm are not
unimodal. For example,
1
1
1
− z f (z) − 1 + 2 f (z)
qm (z) =
2
z
z
so that qm (1) = 0 and
qm (z) =
1
2
1
1
2f (z)
− z f (z) − 2 1 + 2 f (z) +
z
z
z3
so that qm (1) = −2f (1) > 0: qm has an antimode at z = 1.
Remark 2.9. Now, define Ym = log Zm . By the definition of log-symmetry, the
distribution of Ym , with density
pm (x) = 12 (1 − e2x )f (ex ),
x ∈ R,
(2.12)
is symmetric. However, a similar analysis shows that, under the same
assumptions, pm also has an antimode at its centre, x = 0. It seems that typically
qm and pm are bimodal densities.
(c) Distributions for Z specified
We have seen that a smooth R-symmetric unimodal random variable X can
be expressed in terms of a uniform random variable U on (0,1) and a random
variable Z . In this section, we explore some examples of distributions for Z and
their consequences for the distribution of X . They are somewhat instructive in the
sense that a distribution for Z on the interval (0, 1) will produce an R-symmetric
distribution, but not necessarily a smooth one.
Example 2.10. If Z is uniform on (0,1), then the distribution of X is given by
⎧
⎨− 12 log(1 − x 2 ),
if 0 < x < 1,
(2.13)
fU (x) =
⎩log x − 1 log(x 2 − 1), if x ≥ 1.
2
Of course, this density tends to ∞ as x → 1. Few other choices for q are so
accommodating and most result in rather nasty special function densities for f .
This goes, for example, in general, for beta densities. (We shall not write out any
such rebarbative formulae.)
Example 2.11. Formula (2.1) can be written as
−(1/2) log(1−M 2 (x)) q( 1 − e−2w ) dw,
f (x) =
(2.14)
0
where M (x) = min(x, 1/x). Candidates for simple results are therefore densities
that can be written as simple functions of (1 − z 2 ). For example, truncate a
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2089
R-symmetry of probability distributions
p=6
p=5
p=4
p=3
p=2
p=1
fp(x)
0.6
0.4
0.2
0
0
2
4
6
8
10
x
Figure 3. Pdfs fp for p = 1, 2, 3, 4, 5, 6.
symmetric beta distribution on [−1, 1], with parameter p > 0 say, on to support
[0, 1]. This has density
1
22(p−1) B(p, p)
(1 − z 2 )p−1 =
2
(1 − z 2 )p−1 ,
B((1/2), p)
0 < z < 1.
The resulting unimodal R-symmetric density for p = 1 is
fp (x) =
1
{1 − (1 − M 2 (x))p−1 },
(p − 1)B((1/2), p)
(2.15)
where M (x) is again min(x, 1/x). (As p → 1, fp (x) → fU (x) above.) For integer
p > 1, fp (1) is finite and fp is p − 2 times continuously differentiable at 1.
Figure 3 gives plots of fp (x) for various values of p. This demonstrates the
degree of smoothness of f at its mode, and it is seen that as p → 1 the density
tends to ∞ for x → 1.
(d) Unique representation of R-symmetry
It is tempting to consider random variables X as in theorem 2.1 by replacing U
by a random variable V that is not uniform. One such is to take V ∼ beta(n, 1)
for n a positive integer. This is the analogue in the ordinary unimodal case of the
‘a-unimodality’ as defined by Olshen & Savage (1970); see also Dharmadhikari &
Joag-Dev (1988) and Bertin et al. (1997). The pdf of X as an extension of equation
(2.4) in this case is
M (x)
(x − z)n−1
q(z) dz.
(2.16)
f (x) = n
(1/z − z)n
0
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2090
Y. P. Chaubey et al.
However, we can show that only the uniform distribution for U preserves the
R-symmetry. To prove this, note that for Z ∼ q and V ∼ h, for arbitrary densities
q and h on (0, 1), the density of X = V /Z + (1 − V )Z is given by
f (x) =
M (x)
0
1
h
1/z − z
x −z
q(z) dz.
1/z − z
For f to be R-symmetric, we would also need
M (x)
1/x − z
1
1
h
q(z) dz
=
f
1/z − z
1/z − z
x
0
(2.17)
(2.18)
to equal f (x). This is clearly so if h is uniform and clearly not so otherwise,
as the following simple argument shows. Fix on any non-uniform h, such as
beta(n, 1), then for equation (2.17) to hold, the appropriate choice of q depends
on f . On the other hand, any q that satisfies equality of the right-hand sides
of equations (2.17) and (2.18) would not depend on f . Thus, it shows that
(unimodal) R-symmetric X s cannot satisfy X = V /Z + (1 − V )Z , V and Z
defined on (0, 1) and independent, for any V other than uniform.
3. Symmetric distributions associated with R-symmetric distributions
(a) A general representation of R-symmetric distributions
Following Boros & Moll (2004, §13.2), Baker (2008) offers the following result:
∞
∞ b 2
f
cx −
f (u 2 ) du, c > 0, b ≥ 0,
(3.1)
dx =
c
cx
0
0
as a version of the classical, but not so well-known, Cauchy–Schlömilch transformation (for which see Amdeberhan et al. submitted). The original transformation was used to evaluate seemingly intractable integrals; the version at equation
(3.1) is useful for applications in probability and statistics. Thus, it follows that
if g(u), u ≥ 0 is a pdf, termed the mother pdf by Baker (2008), then f (x) =
cg(|cx − b/cx|) is also a pdf√
for x ≥ 0, referred to as the daughter pdf. Moreover,
f is R-symmetric about q = b/c. Note that there is a one-to-one correspondence
between the daughter pdf f and the mother pdf g, f being obtained from g by
shifting and redistributing its probability mass. The following theorem clarifies
the correspondence between the symmetric distributions on the real line and
the R-symmetric distributions with non-negative real support. Without
√ loss of
generality, set the scale parameter c = 1 and the centre of R-symmetry b = 1.
Theorem 3.1. If g(x) is the pdf of a symmetric real-valued random variable X
defined on R, i.e. g(x) = g(−x) for all x ∈ R, then
1
, x > 0,
(3.2)
f (x) = 2g x −
x
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
R-symmetry of probability distributions
2091
is an R-symmetric density and conversely, any R-symmetric density f gives rise
to an ordinary symmetric g on R through
g(x) = f (x + 1 + x 2 ).
(3.3)
(This neat form actually corresponds to a rescaling of g relative to equation (3.2).)
Proof. Let g be the density of an ordinary symmetric distribution on R, then
R-symmetry of f is obvious. The fact that f is necessarily a density can be derived
from equation (3.1). Explicitly,
∞
∞ ∞
1
1
1
f (x) dx =
g x−
g
dx +
− w dw
2
x
w
0
0
0 w
∞ 1
1
=
g x−
1 + 2 dx
x
x
0
∞
g(z) dz = 1.
=
−∞
Here, the substitutions w = 1/x and then z = x − 1/x were used. The converse
follows similarly.
Remark 3.2. The density f defined in equation (3.2) is unimodal if and only if
g is unimodal.
Below are given some examples of R-symmetric densities obtained from some
familiar choices of g.
Example 3.3. Normal g gives rise, in this way, to the RRIG f , equation (1.9).
Example 3.4. A suitably scaled t density on n degrees of freedom has density
proportional to (2 + x 2 )−(n+1)/2 , and gives rise to a certain scaled ‘generalized F ’
R-symmetric density with density proportional to x n+1 /(1 + x 4 )(n+1)/2 (this is the
density of a suitably scaled Fn/2+1,n/2 random variable raised to the 1/4 power).
Example 3.5. The symmetric version of the hyperbolic distribution of
Barndorff-Nielsen (1977) on R suitably scaled has density
1
g(x) =
exp(−x 4 + x 2 )
(3.4)
4K1 (x)
for x > 0 and K1 a Bessel function (see also Barndorff-Nielsen & Blaesild 1983).
Its R-symmetric counterpart from equation (3.2) is the positive hyperbolic
distribution with its two parameters equal to x (see also Barndorff-Nielsen &
Blaesild 1983); it has density
1
1
exp −x x +
(3.5)
f (x) =
2K1 (x)
x
on R+ .
Conversely, some examples of not-so-familiar symmetric unimodal densities
obtained from R-symmetric unimodal densities follow.
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2092
Y. P. Chaubey et al.
p=6
p=5
p=4
p=3
p=2
p=1
gp(x)
0.6
0.4
0.2
0
−3
−2
−1
0
x
1
2
3
Figure 4. Pdfs gp for p = 1, 2, 3, 4, 5, 6.
Example 3.6. Lognormal f gives rise, in this way, to the novel symmetric
unimodal density
1
e−m/2
−1
2
exp − {sinh (x)} .
(3.6)
g(x) = √
2m
2pm
Example 3.7. Density (2.13) yields another remarkable novel symmetric
unimodal density, albeit one with an infinite spike at the origin:
g(x) = − 12 log 2|x|( 1 + x 2 − |x|) .
(3.7)
Example 3.8. Better behaved novel symmetric unimodal densities arise in
similar fashion from equation (2.15):
p−1 1
2
gp (x) =
.
(3.8)
1 − 2|x|( 1 + x − |x|)
(p − 1)B((1/2), p)
Figure 4 gives plots of gp (x) for various values of p that clearly show symmetric
densities, which have fatter tails as p becomes larger. Density (3.8) tends to
density (3.7) as p → 1; so density (3.7) is also shown in figure 4.
Remark 3.9. Note that the equivalent transformation of variables relating Y on
R with X on R+ through Y = X − (1/X ) (Jones 2007) is quite different. It relates
to log-symmetric distributions in the sense that if X follows a log-symmetric
distribution on R+ , then Y follows an ordinary symmetric distribution on
R. This—along with various aspects of the development above—is because
X − 1/X = 2 sinh(log X ).
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
R-symmetry of probability distributions
2093
(b) A Khintchine-type match-up between R-symmetric and log-symmetric
distributions
We can now link the general formulation (3.2) with a general form for q. It is
the case that
1
1
f (x) = 2 1 + 2 g x −
,
(3.9)
x
x
so that
2 1
4
.
q(z) = 3 1 − z g z −
z
z
(3.10)
For any such distribution, if Z ∼ q, then Y = 1/Z − Z follows the distribution
with density (y) ≡ −2yg (y), y > 0.
Now, once more let U ∼ U (0, 1) and, independently, Y ∼ (y), y > 0. Then,
R-symmetric (R) and log-symmetric (L) random variables can be represented in
terms of Y as follows.
— We know from theorem 2.1 that for unimodal R-symmetric distributions,
the associated random variable R can be written as
1 1
(3.11)
− Z = { Y 2 + 4 + (2U − 1)Y }.
R=Z +U
Z
2
— It is also the case, however, that Khintchine’s theorem for ordinary
symmetric unimodal distributions (random variable X , say) on the real line
can be written as X = (2U − 1)Y . But a log-symmetric random variable L
is of the form L = eX . It follows that for log-symmetric distributions based
on unimodal g, the associated random variable L can be written as
L = exp{(2U − 1)Y }.
(3.12)
This seemingly provides an intriguing match-up between unimodal R and L
based on unimodal g in terms of different functions of U and Y .
4. Conclusions
The results of §§2 and 3 have given a number of insights into the theoretical role
and consequences of R-symmetry. Section 2 considered in detail a Khintchine–
type theorem for R-symmetric distributions and various important specific
distributions associated with it. Section 3 considered the equivalence between
R-symmetric distributions and distributions composed of ordinary symmetric
distributions whose scale is transformed by the Cauchy–Schlömilch device. This
also provided another intriguing parallel between R- and log-symmetry.
The above are not direct practical consequences of our work but underlie some
important practical questions, most notably that of the role of IG distributions
in statistical practice. In §1b, we stressed the many G–IG analogies that make
the IG distribution such an attractive option for modelling non-negative data.
These analogies have until now retained an air of mystery: they are true, but
why are they true? The central role of the IG distribution in R-symmetry,
through the RRIG distribution, provides an answer; in particular, our one-line
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2094
Y. P. Chaubey et al.
example 3.3 makes the R-symmetry/Cauchy–Schlömilch link between RRIG and
G distributions, which seems to be the real driver of these analogies. We do
not have space to expand on this in detail here, but note, as just one example,
that the maximum entropy characterization of the RRIG distributions follows
directly by combining Shannon’s characterization of the G distribution with the
Cauchy–Schlömilch transformation.
The research of the first author was partially supported from the author’s Discovery grant from
the Natural Sciences and Engineering Research Council of Canada. The authors are very grateful
to the referees for their fair and helpful remarks.
References
Amdeberhan, T., Glasser, M. L., Jones, M. C., Moll, V. H., Posey, R. & Varela, D. Submitted.
The Cauchy–Schlömilch transformation.
Anderson, T. W. 1955 The integral of a symmetric unimodal function over a symmetric convex set
and some probability inequalities. Proc. Amer. Math. Soc. 6, 170–176. (doi:10.2307/2032333)
Baker, R. 2008 Probabilistic applications of the Schlömilch transformation. Commun. Statist.
Theory Meth. 37, 2162–2176. (doi:10.1080/03610920801892014)
Barndorff-Nielsen, O. 1977 Exponentially decreasing distributions for the logarithm of particle size.
Proc. R. Soc. Lond. A 353, 401–419. (doi:10.1098/rspa.1977.0041)
Barndorff-Nielsen, O. & Blaesild, P. 1983 Hyperbolic distributions. In Encyclopedia of statistics,
vol. 3 (eds S. Kotz & N. L. Johnson), pp. 700–707. New York, NY: Wiley.
Berg, C. 1998 From discrete to absolutely continuous solutions of indeterminate moment problems.
Arab J. Math. Sci. 4, 1–18.
Bertin, E. M. J., Cuculescu, I. & Theodorescu, R. 1997 Unimodality of probability measures.
Dordrecht, The Netherlands: Kluwer.
Birnbaum, Z. W. 1948 On random variables with comparable peakedness. Ann. Math. Statist. 19,
76–81. (doi:10.1214/aoms/1177730293)
Boros, G. & Moll, V. H. 2004 Irresistible integrals: symbolics, analysis and experiments in the
evaluation of integrals. Cambridge, UK: Cambridge University Press.
Dalal, S. R. & Mudholkar, G. S. 1988 A conservative test and confidence region for comparing
heteroscedastic regressions. Biometrika 75, 149–152. (doi:10.1093/biomet/75.1.149)
Das Gupta, S. 1976 S-unimodal functions: related inequalities and statistical applications. Sankhyā
38, 301–314.
Das Gupta, S., Anderson, T. W. & Mudholkar, G. S. 1964 Monotonicity of the power
functions of some tests of the multivariate linear hypothesis. Ann. Math. Statist. 35, 200–205.
(doi:10.1214/aoms/1177703742)
Dharmadhikari, S. & Joag-Dev, K. 1988 Unimodality, convexity, and applications. New York, NY:
Academic Press.
Feller, W. 1971 An introduction to probability theory and its applications, vol. 2, 2nd edn. New
York, NY: Wiley.
Folks, J. L. & Chhikara, R. S. 1976 The inverse Gaussian distribution and its statistical
applications—a review (with discussion). J. Roy. Statist. Soc. Ser. B 40, 263–289.
Ghosh, P. 1974 On generalized unimodality. Commun. Statist. 3, 567–590. (doi:10.1080/
03610927408827159)
Gnedenko, B. V. & Kolmogorov, A. N. 1949 Limit distributions for sums of independent random
variables (Predelnye Raspredeleniya dlya Summ Nezavisimykh Sluchainykh Velichin, MoscowLeningrad, Gostekhizdat, 1949.) [In Russian.] [Translated and annotated by K. L. Chung, with
appendices by J. L. Doob & P. L. Hsu, Addison-Wesley, Reading, MA, 1st edn 1954, 2nd edn
1968].
Heyde, C. C. 1983 On a property of the lognormal distribution. J. Roy. Statist. Soc. Ser. B 29,
392–393.
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
R-symmetry of probability distributions
2095
Jones, M. C. 2002 On Khintchine’s theorem and its place in random variate generation. Amer.
Statist. 56, 304–307. (doi:10.1198/000313002588)
Jones, M. C. 2007 Connecting distributions with power tails on the real line, the half line and the
interval. Int. Statist. Rev. 75, 58–69. (doi:10.1111/j.1751-5823.2007.00006.x)
Jones, M. C. 2008 On reciprocal symmetry. J. Statist. Planning Inf. 138, 3039–3043.
(doi:10.1016/j.jspi.2007.11.006)
Jones, M. C. & Arnold, B. C. 2008 Distributions that are both log-symmetric and R-symmetric.
Elec. J. Statist. 2, 1300–1308. (doi:10.1214/08-EJS301)
Khintchine, I. 1938 On unimodal distributions. Izv. Nauchno-Isled. Inst. Mat. Mech. Tomsk. Gos.
Univ. 2, 1–7. [In Russian].
Laha, R. G. 1961 On a class of unimodal distributions. Proc. Amer. Math. Soc. 12, 181–184.
(doi:10.2307/2034305)
Marshall, A. W. & Olkin, I. 1979 Inequalities: theory of majorization and its applications. New
York, NY: Academic Press.
Medgyessy, P. 1967 On a new class of unimodal infinitely divisible distribution functions and related
topics. Studia Sci. Math. Hungar. 2, 441–446.
Mudholkar, G. S. 1965 A class of tests with monotone power functions for two problems
in multivariate statistical analysis. Ann. Math. Statist. 36, 1794–1801. (doi:10.1214/aoms/
1177699808)
Mudholkar, G. S. 1966 The integral of an invariant unimodal function over an invariant convex
set—an inequality and applications. Proc. Amer. Math. Soc. 17, 1327–1333. (doi:10.2307/
2035735)
Mudholkar, G. S. 1969 A generalized monotone character of d.f.’s and moments of statistics from
some well-known populations. Ann. Inst. Statist. Math. 21, 277–285. (doi:10.1007/BF02532255)
Mudholkar, G. S. 1972 G-peakedness comparison for random vectors. Ann. Inst. Statist. Math. 24,
127–135. (doi:10.1007/BF02479744)
Mudholkar, G. S. & Natarajan, R. 2002 The inverse Gaussian models: analogues of
symmetry, skewness and kurtosis. Ann. Inst. Statist. Math. 54, 138–154. (doi:10.1023/
A:1016173923461)
Mudholkar, G. S. & Tian, L. 2002 An entropy characterization of the inverse Gaussian
distribution and related goodness-of-fit test. J. Statist. Planning Inf. 102, 211–221. (doi:10.1016/
S0378-3758(01)00099-4)
Mudholkar, G. S. & Wang, H. 2007 IG-symmetry and R-symmetry: interrelations and applications
to the inverse Gaussian theory. J. Statist. Planning Inf. 137, 3655–3671. (doi:10.1016/
j.jspi.2007.03.041)
Mudholkar, G. S. & Wang, H. 2009 R-Symmetry and the power functions of the tests
for inverse Gaussian means. Commun. Statist. Theory Meth. 38, 2178–2186. (doi:10.1080/
03610920802499496)
Mudholkar, G. S., Wang, H. & Natarajan, R. 2009 The G–IG analogies and robust tests for inverse
Gaussian scale parameters. In Advances in multivariate statistical methods (ed. A. SenGupta),
pp. 285–304. London, UK: World Scientific.
Olshen, R. A. & Savage, L. J. 1970 A generalized unimodality. J. Appl. Prob. 7, 21–34. (doi:10.2307/
3212145)
Pakes, A. G. 1996 Length-biasing and laws equivalent to the log-normal. J. Math. Anal. Appl. 197,
825–854. (doi:10.1006/jmaa.1996.0056)
Rao, C. R. 1965 Linear statistical inference and its applications. New York, NY: Wiley.
Schrödinger, E. 1915 Zur theorie der fall-und steigversuche an teilchenn mit Brownsche bewegung.
Phys. Zeit. 16, 289–295.
Seshadri, V. 1965 On random variables which have the same distribution as their reciprocals.
Canad. Math. Bull. 8, 819–824.
Seshadri, V. 1993 The inverse Gaussian distribution: a case study in exponential families. Oxford,
UK: Clarendon Press.
Shannon, C. E. 1949 The mathematical theory of communication. New York, NY: Wiley.
Shepp, L. A. 1962 Symmetric random walk. Trans. Amer. Math. Soc. 104, 144–153. (doi:10.2307/
1993938)
Proc. R. Soc. A (2010)
Downloaded from http://rspa.royalsocietypublishing.org/ on July 12, 2017
2096
Y. P. Chaubey et al.
Sherman, S. 1955 A theorem on convex sets with applications. Ann. Math. Statist. 26, 763–766.
(doi:10.1214/aoms/1177728435)
Smoluchowski, M. V. 1915 Notiz über die berechning der Brownschen molkularbewegung bei des
Ehrenhaft-millikanchen versuchsanordnung. Phys. Zeit. 16, 318–321.
Sun, T. C. 1967 A note on unimodality of distribution functions of class L. Ann. Math. Stat. 38,
1296–1299. (doi:10.1214/aoms/1177698804)
Tong, Y. L. 1980 Probability inequalities in multivariate distributions. New York, NY: Academic
Press.
Tweedie, M. C. K. 1945 Inverse statistical variates. Nature 155, 453. (doi:10.1038/155453a0)
Vitale, R. A. 1990 The Brunn–Minkowski inequality for random sets. J. Multivariate Anal. 90,
286–293. (doi:10.1016/0047-259X(90)90052-J)
Wintner, A. 1938 Asymptotic distributions and infinite convolutions. Ann Arbor, MI: Edwards
Brothers.
Wolfe, S. J. 1971 On the unimodality of L functions. Ann. Math. Statist. 42, 912–918. (doi:10.1214/
aoms/1177693320)
Proc. R. Soc. A (2010)