Indifference-Zone-Free Selection of the Best

Submitted to Operations Research
manuscript (Please, provide the manuscript number!)
Indifference-Zone-Free Selection of the Best
Weiwei Fan
Department of Management Science, School of Management, University of Science and Technology of China, Hefei, China,
[email protected]
L. Jeff Hong
Department of Economics and Finance and Department of Management Sciences, College of Business, City University of
Hong Kong, Kowloon Tong, Hong Kong, jeff[email protected]
Barry L. Nelson
Department of Industrial Engineering and Management Sciences, Northwestern University, Evanston, Illinois, 60208,
[email protected]
Many procedures have been proposed in the literature to select the simulated alternative with the best
mean performance from a finite set of alternatives. Among these procedures, frequentist procedures are typically designed under either the subset-selection (SS) formulation or the indifference-zone (IZ) formulation.
Both formulations may encounter problems when the goal is to select the unique best alternative for any
configuration of the means. In particular, SS procedures may return a subset that contains more than one
alternative, and IZ procedures hinge upon the relationship between the chosen IZ parameter and the true
mean differences that is unknown to decision makers a priori. In this paper, we propose a new formulation
that guarantees to select the unique best alternative with a user-specified probability of correct selection
(PCS), as long as the means of alternatives are unique, and we design a class of fully sequential procedures
under this formulation. These procedures are parameterized by the PCS value only, and their continuation
boundaries are determined based on the Law of the Iterated Logarithm. Furthermore, we show that users
can add a stopping criterion to these procedures to convert them into IZ procedures, and we argue that these
procedures have several advantages over existing IZ procedures. Lastly, we conduct an extensive numerical
study to show the performance of our procedures and compare their performance to existing procedures.
1. Introduction
Decision makers often face the problem of selecting the best from a finite set of alternatives, where
the best refers to the alternative with the largest (or smallest) mean performance. For instance, an
inventory manager may need to select an inventory policy to minimize the total mean cost, while a
financial investor may want to choose an investment strategy to maximize the mean payoff. In many
practical situations, the mean performances of these alternatives are not explicitly available and
can only be evaluated by running simulation experiments. This is known as the selection-of-thebest problem. To solve this problem, a decision procedure is often needed to determine the proper
sample sizes of all alternatives and to decide which alternative to select. There is a large body of
1
2
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
literature on designing such procedures, see Bechhofer et al. (1995), Kim and Nelson (2006a) and
Branke et al. (2007) for introductions and overviews.
Among these procedures in the literature, there are frequentist and Bayesian approaches. Frequentist procedures (e.g., Rinott 1978, Kim and Nelson 2001) allocate simulation observations to
different alternatives to achieve a pre-specified probability of correct selection (PCS) even for the
least favorable configuration of the means. They are typically conservative (i.e., requiring more
samples than necessary) for an average case. Bayesian procedures (e.g., Chick and Inoue 2001,
Chick and Frazier 2012) often allocate a finite computing budget to different alternatives to either
maximize the posterior (Bayesian) PCS or minimize the opportunity cost. They often require fewer
observations than frequentist procedures, but typically do not provide a guaranteed (frequentist)
PCS. An exception is Frazier (2014) who recently proposes a Bayes-inspired procedure to achieve
a pre-specified frequentist PCS. In this paper we take the frequentist viewpoint and our goal is to
design procedures that deliver a user-specified PCS.
The first frequentist selection-of-the-best formulations were designed in the 1950s, when observations were often collected through physical experiments, e.g., agricultural experiments or clinical
trials. Each of these experiments may take quite a long time (e.g., weeks to months) to conduct,
and it only makes sense to conduct these experiments in batches. Therefore, the stage-wise procedures that select the best at the end of the last stage were prevalent. For these procedures, the
critical issue is to determine the appropriate sample size for each alternative so that the best may
be selected. Notice that the means of the best and the second-best alternatives may be arbitrarily
close. Therefore, guaranteeing to find the unique best for all configurations of the means could
require an arbitrarily large number of observations, which is not practical. To avoid this difficulty,
various formulations were proposed to soften the original goal of selecting the unique best alternative. One is the subset-selection (SS) formulation of Gupta (1956, 1965), which guarantees to select
a random-sized subset of the alternatives that contains the best (instead of selecting the unique
best). The procedures of Panchapakesan et al. (1971) and Sullivan and Wilson (1989) are also
designed under this formulation. The other formulation is the indifference-zone (IZ) formulation
of Bechhofer (1954), which guarantees to select the unique best alternative when the difference
between the means of the best and the second-best is at least δ, where δ > 0 is called an IZ parameter. The procedures proposed by Rinott (1978), Clark and Yang (1986) and many others are also
designed under this formulation.
Starting in the 1990s there was a paradigm change in selection-of-the-best problems. Under this
new paradigm, observations of alternatives are often obtained through computer simulation experiments. Each of these experiments takes a fraction of a second to a few minutes and they are often
generated sequentially from either a single processor or a small number of parallel processors. This
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
3
sequential nature of computer simulation experiments motivates the use of sequential procedures
that date back to Paulson (1964) under the IZ formulation. The major milestone of this stream of
research was the paper of Kim and Nelson (2001) that completely adapts fully sequential procedures
to computer simulation experiments. These fully sequential procedures typically specify a continuation region. They then approximate the sum of differences between two competing alternatives
as a Brownian motion (BM) process with drift, and make an elimination decision when the process exits the continuation region. Different from stage-wise procedures, fully sequential procedures
gradually gather information (i.e., simulation observations) on the unknown differences between
the means of competing alternatives and eliminate the inferior alternatives once enough statistical
evidence is available. Therefore, they tend to be more efficient, i.e., need fewer observations, than
their stage-wise counterparts.
Even though many fully sequential frequentist procedures have been proposed since the work of
Kim and Nelson (2001), e.g., the procedures of Batur and Kim (2006), Hong and Nelson (2005,
2007), Pichitlamken et al. (2006), Hong (2006), Kim and Nelson (2006b) and many others, none of
these questioned the necessity of the IZ formulation. Unlike stage-wise procedures where information on the means of the competing alternatives are unknown, fully sequential procedures gather
this information as they progress. This information may free us from having to provide an IZ
parameter and help us to select the unique best alternative for any configuration of the means.
In this paper we propose a new formulation that selects the best alternative with a user-specified
PCS, as long as all the alternatives have unique means. Under this formulation, the means of any
pair of alternatives are not equal and can be arbitrarily close. Therefore, the key to designing such
a procedure is to construct a continuation region where, with the specified PCS, no elimination
decision will be made if the mean difference between these two alternatives is zero, while a correct
elimination decision will be made if it is not. To achieve this, we need to construct a continuation
region so that a BM with no drift will stay inside while a BM with drift will exit from the correct
side. Notice that BM with and without drift differ in the rates at which they grow. In particular, a
BM with positive drift approaches infinity at the rate O(t), while a BM without drift approaches
√
infinity at a rate bounded by O( t log log t ) due to the Law of the Iterated Logarithm (Durrett
2010). In light of this, any continuation region, formed by boundaries which grow at a rate between
√
O( t log log t ) and O(t), can achieve our objective.
To further determine the design parameters of such continuation regions, we need to be able
to evaluate the first-exit probability of a BM from the given continuation region. However, it
is generally difficult or even impossible to explicitly analyze the first-exit-time distribution of a
BM (see Durbin 1985). To solve the problem, we consider two approaches. In the first approach,
we obtain the design parameter by numerically solving a one-dimensional stochastic root-finding
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
4
problem, which can be solved efficiently with the use of an importance sampling scheme. In the
second approach, we obtain the design parameter by evaluating the first-exit probability using an
asymptotic result and we prove the validity of the procedures as PCS goes to 1. It is worthwhile
noting that the asymptotic regime of letting PCS go to 1 is a classical one and it dates back to
Perng (1969) and Dudewicz (1969).
Beyond selecting the unique best alternative, our procedures can also be used as sequential
subset-selection procedures. In particular, at any time before the procedures terminate, the surviving subset of alternatives has retained the best alternative with the same probability guarantee.
One shortcoming of our procedures is that they may not stop when two or more alternatives
have the same means. To avoid this situation, we propose a stopping criterion for our procedures.
The stopping criterion requires the user to specify an error tolerance ε. We show that, with the
same PCS, the selected alternative is either the best or within ε of the best. This indicates that
our procedures provide both the correct selection guarantee and good selection guarantee, which
is the goal of Ni et al. (2014). If one treats ε as an IZ parameter, then our procedures become
IZ procedures. However, unlike IZ procedures, the average sample size of our procedures is typically insensitive to the setting of the error tolerance, thus one may set it very small for practical
implementations.
Our research is closely related to the literature on best-arm identification in stochastic multiarmed bandit (MAB) problems. (Readers interested in the MAB problems may refer to Bubeck
and Cesa-Bianchi (2012) for a comprehensive overview). The best-arm identification procedures
seek the same goal as our procedures do, i.e., selecting the unique best arm with a pre-determined
probability. In contrast, our procedures and these best-arm identification procedures differ in at
least three aspects. Firstly, they employ different elimination mechanisms. In particular, the bestarm identification procedures (e.g. Karnin et al. 2013, Jamieson and Nowak 2014) sequentially
construct a confidence bound for the mean reward of each arm based on the estimated sample
mean, and make eliminations if there are non-overlapping confidence bounds. Secondly, the bestarm identification procedures assume the cumulant generating function of each reward is bounded
by a known convex function, while our procedures only require a finite joint moment generating
function of rewards in a neighborhood of zero vector (see Theorem 2) and it need not be known.
Lastly, our procedures are able to allow dependence among the rewards of different arms and thus
common random numbers can be used to speed up the selection of the best.
The rest of our paper is organized as follows. In Section 2, we introduce a new selection-ofthe-best formulation. In Section 3, we design a class of fully sequential IZ-free procedures under
the new formulation when observations from all alternatives are jointly normally distributed. In
Section 4, we relax the normality assumption to design the corresponding sequential procedures
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
5
that are asymptotically valid. In Section 5 we analyze the asymptotic efficiency of our procedures
and discuss a stopping criterion in Section 6. In Section 7 we conduct a comprehensive numerical
study to understand our procedures and compare them to existing procedures, and we conclude in
Section 8.
2. A New Selection-of-the-Best Formulation
Suppose there are k (≥ 2) competing alternatives at the beginning of the selection process, and
the goal is to select the alternative that has the largest mean performance. For i = 1, 2, . . . , k,
denote µi as the unknown mean performance of alternative i and we evaluate it through computer
simulation. Without loss of generality, assume µ1 > µ2 > · · · > µk , implying that the best alternative
is unique and alternative 1 is the best. It is worthwhile noting that we make no assumption on
the difference between µ1 and µ2 because there is often no such knowledge in practice. Further,
let 1 − α (0 < α ≤ 1 − 1/k) denote the user-specified PCS. In other words, users target at selecting
alternative 1 with probability at least 1 − α.
To achieve this target, there are two different frequentist formulations in the literature. They
are the SS and IZ formulations. As stated in Section 1, both of them may encounter difficulties.
In light of this, we propose a new selection-of-the-best formulation in this section. Under the new
formulation, procedures guarantee to select the best alternative with a user-specified PCS value
1 − α, as long as their means are unique, i.e., µ1 > µ2 > · · · > µk . Equivalently, procedures under
the new formulation guarantee to satisfy
P{select alternative 1} ≥ 1 − α.
(1)
Therefore, to use the procedures under the new formulation, one only needs to specify the PCS
value 1 − α. Compared to the IZ formulation, the new formulation frees users from having to specify
an IZ parameter, and compared to the SS formulation, it selects the unique best alternative instead
of a random subset, but can select a subset if procedures are terminated early.
In the following two sections, we design sequential procedures under the new formulation for
cases where observations are jointly normally distributed as well as cases where observations are
generally distributed, and prove that these procedures can deliver the required statistical guarantee,
i.e., Equation (1). We emphasize that we use sequential procedures not only because they are more
efficient but also because stage-wise procedures cannot achieve the necessary statistical guarantee.
3. Procedures with Normally Distributed Observations
Let Xir , r = 1, 2, . . . , denote the rth independent observation from alternative i, for i = 1, 2, . . . , k.
In this section, we consider the case where (X1r , X2r , . . . , Xkr )(r = 1, 2, . . .) are jointly normally
6
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
distributed with unknown mean vector (µ1 , µ2 , . . . , µk ) and unknown positive definite covariance
matrix. The assumption of possible dependence among observations from different alternatives
enables the use of common random numbers (CRNs). The task of this section is to design a
sequential procedure that deliver the statistical guarantee stated in (1).
Existing sequential procedures, such as the KN procedure, often decompose the selection-of-thebest problem into pairwise comparisons. In each pairwise comparison, they then approximate the
partial-sum difference process between two competing alternatives by a Brownian motion with drift
and design a continuation region to eliminate the inferior alternative based on the first-exit time of
the Brownian motion from it. In this paper, we adopt the same mechanism to design a sequential
procedure under the new formulation and, consequently, we need to address the following two
issues: (a) constructing a Brownian approximation to the partial-sum difference process, and (b)
designing a proper continuation region.
Under the normality assumption, it is straightforward to construct a Brownian approximation.
Consider a pairwise comparison between alternatives i and j, where i ̸= j. Let X̄i (n) denote the
sample mean calculated from the first n observations from alternative i and define
2
tij (n) = n/σij
, and Zij (tij (n)) = tij (n)[X̄i (n) − X̄j (n)],
(2)
2
where σij
= Var[Xi1 − Xj1 ]. Throughout this paper, we define B(·) as a Brownian motion (BM)
with unit variance and no drift, and define B∆ (·) as a BM with unit variance and drift ∆. Theorem
1 of Hong (2006) shows that the random sequences {Zij (tij (n)) : n ≥ 1} and {Bµi −µj (tij (n)) : n ≥ 1}
have the same joint distribution. Therefore, {Zij (tij (n)) : n ≥ 1} can be viewed as a BM with drift
µi − µj observed at the discrete time points {tij (1), tij (2), . . .}.
Let (−gc (t), gc (t)) denote the continuation region used in the procedure, where c is a constant
determined based on the number of alternatives k, the PCS value 1 − α and the common first-stage
sample size n0 . As discussed in Section 1, the continuation region is expected to satisfy that, with
a pre-specified PCS, a BM with no drift stays inside, while a BM with drift exits from the correct
side. Notice that the former grows to infinity at a rate slower than that of the latter. In particular,
the Law of the Iterated Logarithm (Durrett 2010) states that, the rate of the former is bounded
√
by O( t log log t ) while the rate of the latter is O(t). The rate difference motivates the use of a
boundary function that increases at rates in-between to construct the targeted continuation region.
In light of this, boundaries of interest are categorized into the following set,
{
}
gc (t)
gc (t)
1
G = gc (t) ∈ C [0, ∞) : lim
= 0, and lim inf √
∈ (0, ∞] ,
t→∞
t→∞
t
t log log t
where C 1 [0, ∞) denotes the set of functions whose first derivatives are continuous on [0, ∞). Furthermore, for the ease of theoretical analysis and practical implementation, we choose gc (t) from
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
7
G such that gc (t) is increasing in t and gc (t)/t is decreasing in t for any fixed c. For instance, we
√
√
can choose 2t log(c + log(t + 1)) (c ≥ 1), [c + log(t + 1)](t + 1) (c > 0) or ctβ (1/2 < β < 1, c > 0).
The determination of the design parameter c is deferred to Section 3.1.
3.1. The Procedure
In this subsection, we propose a sequential procedure designed based on the key idea stated above.
Particularly, the procedure uses a first stage to estimate the unknown variances and determine a
proper continuation region, and then collects observations sequentially to eliminate inferior alternatives until only one is left.
Procedure 1 (Sequential Procedure for Normally Distributed Observations)
Setup. Select the PCS 1 − α (0 < α ≤ 1 − 1/k), and a common first-stage sample size n0 ≥ 2.
Choose a boundary gc (t) from G and calculate the constant c that is the unique root to the following
equation,
}]
[ {
(
)
α
n0 − 1
Y
=
E P B(T ) ≤ −gc
T
, T < ∞Y
,
(3)
Y
n0 − 1
k−1
}
{
(
)
where T = inf t : |B(t)| ≥ gc n0Y−1 t n0Y−1 and Y is a chi-squared distributed random variable with
n0 − 1 degrees of freedom that is independent of B(·). Please refer to Appendix B on how to find
the root c of Equation (3).
Initialization. Let I = {1, 2, . . . , k } be the set of alternatives in contention. For each i ∈ I, simulate n0 observations Xi1 , Xi2 , . . . , Xin0 from alternative i and calculate its sample mean as
0
1 ∑
Xil .
n0 l=1
n
X̄i (n0 ) =
For any i, j ∈ I with i ̸= j, calculate the sample variance of their difference as
0
[
]2
1 ∑
=
Xir − Xjr − (X̄i (n0 ) − X̄j (n0 )) .
n0 − 1 r=1
n
2
Sij
Set n = n0 .
Screening. Let I old = I and for any i, j ∈ I with i ̸= j, let
2
τij (n) = n/Sij
Let
and
Zij (τij (n)) = τij (n)[X̄i (n) − X̄j (n)].
{
}
I = I old \ i ∈ I old : Zij (τij (n)) ≤ −gc (τij (n)) for some other j ∈ I old .
Stopping Rule. If |I | = 1, stop and select the alternative whose index is in I as the best. Otherwise,
take one additional observation from each alternative i ∈ I, set n = n + 1 and go to Screening.
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
8
Remark 1. Procedure 1 can work with many choices of boundaries (i.e., G ). In the numerical
√
studies reported in Section 7, we use a specific boundary gc (t) = [c + log(t + 1)](t + 1) as an
instance.
Procedure 1, like those of Kim and Nelson (2001) and Hong (2006), is essentially a sequential
screening procedure and works as follows. It includes all the alternatives into I at the Initiation
step and eliminates the inferior alternatives from I over the Screening steps until only one alternative is left. At each Screening step, it considers all the possible pairwise comparisons between
the surviving alternatives and eliminates the inferior alternatives based on the first-exit time of the
partial-sum difference process from the given continuation region. Take the comparison between
alternatives i and j for an instance. When the partial-sum difference process Zij (τij (·)) exits the
continuation region from below (above), Procedure 1 eliminates alternative i (j), otherwise it eliminates neither of them and continues collecting observations (see Figure 1). Once an alternative is
eliminated, it will not be considered again.
Z ij(τ ij(n))
eliminate alternative j
g (τ (n))
c ij
τ (n)
ij
eliminate alternative i
c
Figure 1
(τ (n))
ij
Continuation region of our sequential procedures. (The x-axis is well defined because τij (n) increases
linearly with n.)
Even though Procedure 1 is designed to select the unique best alternative, users can stop Procedure 1 at any time before termination and it still guarantees to have retained the best alternative.
This result is summarized as the following Lemma 1.
Lemma 1. Define N as the number of replications when Procedure 1 terminates and only one
alternative is left. Let I(n) for n ≤ N denote the set of surviving alternatives in Procedure 1 after
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
9
n replications. If Procedure 1 delivers the statistical guarantee as stated in Equation (1), then we
have, for any n ∈ Z + ,
P{1 ∈ I(n ∧ N )} ≥ 1 − α,
where x ∧ y denotes the smaller of x and y.
Lemma 1 shows that, if there is a maximal stopping time n, Procedure 1 guarantees with at
least probability 1 − α to either select the best alternative before n or retain the best alternative
in a subset when it is stopped at n. Therefore, Procedure 1 can be utilized as a sequential subsetselection procedure when users have a sampling budget on screening. Besides, Procedure 1 tends
to be more efficient than the existing subset-selection procedures (e.g., Gupta 1956) that use equal
allocation of observations among alternatives, especially when the number of alternatives is large.
This arises because Procedure 1 tends to allocate fewer observations to the significantly inferior
alternatives.
3.2. Statistical Validity
In this subsection, we establish the statistical validity of Procedure 1, i.e., it delivers the statistical
guarantee stated in Equation (1). To achieve this goal, we need several lemmas.
In each pairwise comparison, an incorrect-selection event occurs if the better alternative is eliminated, or equivalently, the partial-sum difference process exits the continuation region from the
wrong direction. Notice that we only collect observations at discrete time points in the procedure.
Hence, the partial-sum process corresponds to observing a BM at discrete time points. The following
lemma states that, under very general conditions, the procedure designed for the continuous-time
BM provides an upper bound on the probability of incorrect selection (PICS) for the discrete-time
BM.
Lemma 2. Let B∆ (t) denote a continuous-time Brownian motion with drift ∆ > 0. A discrete-time
process is obtained by observing B∆ (t) at a (possibly) random, increasing sequence of times {ti : i =
1, 2, . . .} taking values in a given countable set. The value of ti depends on B∆ (t) only through its
values in the period [0, ti−1 ]. Define Td = min{ti : |B∆ (ti )| ≥ g(ti )} and Tc = inf {t : |B∆ (t)| ≥ g(t)},
where g(t) ∈ G . Then we have
(1) Tc ≤ Td < ∞, w.p.1,
(2) P{B∆ (Td ) ≤ −g(Td )} ≤ P{B∆ (Tc ) ≤ −g(Tc )}. (Jennison et al. 1980)
The drift of the BM refers to the difference of two competing means, which is unknown in our
setting. In the following lemma, we show that PICS of the BM with the unknown drift is bounded
by its counterpart without drift. This statement is valid for general stochastic processes, not just
for BM.
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
10
Lemma 3. Let Z(·) denote a (discrete-time or continuous-time) stochastic process. (Assume that
Z(·) has right-continuous sample paths if Z(·) is a continuous-time stochastic process.) Let Z∆ (t) =
Z(t) + ∆t, where ∆ > 0. Further, define T and T∆ as the stopping times that Z(·) and Z∆ (·) first
exit the region (−g(t), g(t)), where g(t) is a continuous and non-negative function. Then, we have
that
P{Z∆ (T∆ ) ≤ −g(T∆ ), T∆ < ∞} ≤ P{Z(T ) ≤ −g(T ), T < ∞}.
Now we can establish the statistical validity of Procedure 1 and summarize it as Theorem 1.
Theorem 1. Suppose that there are k alternatives and we are interested in selecting the alternative
with the largest mean. Assume that (X1r , X2r , . . . , Xkr ), r = 1, 2, . . ., are independent and identically
distributed (i.i.d.) and jointly normally distributed with unknown mean vector (µ1 , µ2 , . . . , µk ) and
unknown positive definite covariance matrix. If µ1 > µ2 > . . . > µk , Procedure 1 stops in finite time
with probability 1 (w.p.1) and selects the best alternative (i.e., alternative 1) with probability at
least 1 − α, i.e., P{select alternative 1} ≥ 1 − α.
Remark 2. Theorem 1 only requires the covariance matrix to be positive definite. Therefore, it
allows the use of CRNs that introduces positive correlation among observations from different
alternatives to make the comparisons sharper.
Proof. For any pair of alternatives i and j with i ̸= j, since µi − µj ̸= 0, we see that the elimination
between them occurs in finite time w.p.1 by part (1) of Lemma 2. Similarly, we have all the pairwise
eliminations occur in finite time w.p.1. Therefore, Procedure 1 stops in finite time w.p.1.
When Procedure 1 stops, define an incorrect-selection (ICS) event happens when alternative 1
is eliminated by some other alternative. Then, we have
{k
}
∪
P{ICS} = P
{alternative 1 is eliminated by alternative i}
i=2
≤
k
∑
P{alternative 1 is eliminated by alternative i}.
(4)
i=2
The inequality in (4) arises from the Bonferroni inequality. For any i = 2, 3, . . . , k, we denote ICSi as
the incorrect-selection event that alternative 1 is eliminated by alternative i. To prove the statistical
validity, we view each incorrect-selection event ICSi separately. In particular, we have
P{ICSi } = P {Z1i (τ1i (Ni )) ≤ −gc (τ1i (Ni )), Ni < ∞} ,
{
}
2
2
and it follows
where Ni = min n ≥ n0 : Z1i (τ1i (n)) ≥ gc (τ1i (n)), n ∈ Z + . Let Y = (n0 − 1)S1i
/σ1i
that
{
( 2
) 2
}
σ1i
S1i
P{ICSi } = P Z1i (t1i (Ni )) ≤ −gc
t
(N
)
,
N
<
∞
i
i
2 1i
2
S1i
σ1i
}]
[ {
(
)
n0 − 1
Y
= E P Z1i (t1i (Ni )) ≤ −gc
t1i (Ni )
, Ni < ∞Y
.
Y
n0 − 1
(5)
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
11
Under the normality assumption, we have shown {Z1i (t1i (n)) : n ∈ Z} can be viewed as a BM with
drift µ1 − µi observed at the discrete time points {t1i (n) : n ≥ n0 }. By Basu’s theorem (Basu 1955),
we see Y follows a chi-squared distribution with n0 − 1 degrees of freedom and is independent of
Z1i (t1i (n0 )) = t1i (n0 )[X̄1 (n0 ) − X̄i (n0 )]. Further, Y is a function of X11 − Xi1 , X12 − Xi2 , . . . , X1n0 −
Xin0 and thus is independent of X1,n0 +1 − Xi,n0 +1 , X1,n0 +2 − Xi,n0 +2 , . . .. Therefore, Y is independent
of Z1i (t1i (n)) for any n ≥ n0 . This implies that, conditioning on Y , {Z1i (t1i (n)) : n ≥ n0 } can still
be viewed as a BM with drift µ1 − µi observed at the discrete time points {t1i (n) : n ≥ n0 }. Then,
we get that
}]
)
[ {
(
Y
n0 − 1
Ti
, Ti < ∞Y
by Lemma 2
(5) ≤ E P Bµ1 −µi (Ti ) ≤ −gc
Y
n0 − 1
)
}]
[ {
(
n0 − 1
Y
≤ E P B(T ) ≤ −gc
T
, T < ∞Y
by Lemma 3
Y
n0 − 1
α
=
,
k−1
}
{
}
{
(
(
)
)
where Ti = inf t : |Bµ1 −µi (t)| ≥ gc n0Y−1 t n0Y−1 and T = inf t : |B(t)| ≥ gc n0Y−1 t n0Y−1 . The last
equation holds due to the choice of the design parameter c (see Equation (3)). Furthermore, combining with (4), we have that P{ICS} ≤ α. Therefore, the conclusion of this theorem holds.
2
In Theorem 1, we established the statistical validity of Procedure 1 for the case when all the
means are unique, i.e., µ1 > µ2 > . . . > µk . More generally, the best alternative may be unique and
there might be potential ties amongst the non-optimal alternatives, i.e., µ1 > µ2 ≥ . . . ≥ µk . For
this case, we show in the following proposition that Procedure 1 can provide a statistical guarantee
that is similar to but weaker than that of Theorem 1.
Proposition 1. Under the same assumptions as in Theorem 1, if µ1 > µ2 ≥ . . . ≥ µk ,
Procedure 1 terminates and selects alternative 1 with probability at least 1 − α, i.e.,
P{termiante and select alternative 1} ≥ 1 − α.
Proof. For any sample path, part (1) of Lemma 2 implies that Procedure 1 does not terminate in
finite time only if all the remaining alternatives have the same mean. Because the best alternative
is unique, it suffices to state that the procedure does not terminate in finite time only if alternative
1 is eliminated. Besides, we have
{
P{alternative 1 is eliminated} = P
k
∪
}
{alernative 1 is elimianted by alternative i}
i=2
≤
k
∑
i=2
P{alernative 1 is elimianted by alternative i}.
(6)
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
12
The inequality above arises from the Bonferroni inequality. Notice that (6) is the same as (4),
which is shown to be no larger than α in the proof of Theorem 1. Therefore, we have
P{Procedure 1 does not terminate in finite time} ≤ P{alternative 1 is eliminated} ≤ α,
and
P{terminate and select alterantive 1} = 1 − P{alternative 1 is eliminated} ≥ 1 − α.
2
Therefore, the proposition holds.
If there are tied means among alternatives other than the best, it is possible that Procedure 1
may not terminate in finite time. Proposition 1 proves that the probability is at most α. Therefore,
in practical situations when users are unsure whether the means are unique, the Procedure with a
stopping criterion is recommended to use and the detailed discussion is deferred to Section 6.
4. Procedures with Generally Distributed Observations
In practice, observations are rarely jointly normally distributed, and may even follow different distributions across different alternatives. Therefore, we are motivated to relax the normality assumption
and extend our procedure to solve general selection-of-the-best problems.
When observations are generally distributed, it appears difficult to design procedures that deliver
a finite-sample statistical validity as we do in Section 3. Therefore, we propose a new procedure that
delivers an asymptotic statistical validity, i.e., the procedure satisfies Equation (1) in a meaningful
limit.
Recall that, to design a sequential procedure, we face two issues: (a) constructing a Brownian
approximation to the partial-sum difference process, and (b) designing a proper continuation region.
For the case of generally distributed observations, neither of these two issues are trivial.
Consider a pairwise comparison between alternatives 1 and i with i ̸= 1. As in Equation (2),
let Z1i (·) denote the partial-sum difference process on a set of time points {t1i (n) : n = 1, 2, . . .}.
Under the normality assumption, we have shown in Section 3 that {Z1i (t1i (n)) : n = 1, 2, . . .} can be
viewed as a Brownian motion process observed at discrete time points. However, this will not be
true for a general case. Therefore, we consider a standardized version of {Z1i (t1i (n)) : n = 1, 2, . . .}
with a scaling parameter M , that is
⌊M
∑s⌋
Z1i (t1i (⌊M s⌋)) − t1i (M s)(µ1 − µi )
√
C1i (M, t1i (s)) :=
=
M
l=1
(X1l − Xil ) − M s(µ1 − µi )
√
, for s > 0(7)
2
M σ1i
where ⌊x⌋ denotes the largest integer that is no bigger than x. Then we approximate (7) by a
BM in the limit using the following Functional Central Limit Theorem (see, e.g., Theorem 4.3.2 in
Whitt 2002a).
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
13
Lemma 4 (Functional Central Limit Theorem). Suppose that {Xi : i ≥ 1} is a sequence of
i.i.d. random variables with mean µ and finite variance σ 2 . Then,
∑⌊nt⌋
i=1 Xi − ntµ d
√
→ B(t), as n → ∞, in D[0, ∞),
nσ
d
where → denotes convergence in distribution and D is defined as the Skorohod space.
According to Lemma 4 and the self-similarity scaling property of a BM, we have that
−1
C1i (M, t1i (s)) → σ1i
B(s) = B(t1i (s)), as M → ∞, in D[0, ∞),
d
d
(8)
d
where X = Y denotes that X and Y are identically distributed. Even though C1i (M, t1i (s)) is
defined for a general distribution, Lemma 4 guarantees its convergence to a known BM in the limit.
We call the parameter M the scaling parameter because it determines how the partial-sum
process Zi1 (t1i (n)) is scaled and how fast the scaled process Ci1 (M, t1i (s)) converges to a BM. To
ensure the convergence in (8), we only need M to grow to infinity.
We address the second issue by choosing one boundary gc (t) from the set G , similar to that
in Section 3. To determine its design parameter c, we need to evaluate the first-exit probability,
which shrinks to zero in the asymptotic regime. In this situation, it becomes impractical to find
c numerically as in Appendix B. Fortunately, there is a body of literature on asymptotic approximations of this type of first-exit probabilities. For instance, Jennen and Lerche (1981) provide the
asymptotic approximations for a list of boundaries. In the following lemma, we present one of these
√
approximations where the boundary is gc (t) = [c + log(t + 1)](t + 1). In this lemma, we need to
use the concept of asymptotic equivalence. In particular, we say f (c) is asymptotically equivalent
c→∞
to g(c) with respect to c, denoted as, f (c) ∼ g(c), if limc→∞ f (c)/g(c) = 1.
√
Lemma 5 (Jennen and Lerche (1981)). Let gc (t) = [c + log(t + 1)](t + 1) (c ≥ 0), a nonnegative continuous function, and let B(t) be a Brownian motion without drift. Then P{B(T ) ≤
c→∞ 1 −c/2
e
,
2
−gc (T ), T < ∞} ∼
√
where T = inf {t : |B(t)| ≥ gc (t)}.
Lemma 5 provides a closed-form approximation to the first-exit probability when gc (t) =
(c + log(t + 1))(t + 1). This allows us to choose the value of c more easily compared to the root-
finding problem (3) in Section 3, although this value is only an approximation.
4.1. The Procedure
In this subsection, we propose a sequential procedure for generally distributed observations, i.e.,
Procedure 2.
Procedure 2 (Sequential Procedure for Generally Distributed Observations)
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
14
Setup. Select the PCS 1 − α (0 < α ≤ 1 − 1/k), and a common first-stage sample size n0 ≥ 2.
√
Choose the boundary as gc (t) = [c + log(t + 1)](t + 1) and set c = −2 log(2α/(k − 1)).
Initialization. Let I = {1, 2, . . . , k } be the set of alternatives in contention. For all i = 1, 2, . . . , k,
simulate n0 observations Xi1 , Xi2 , . . . , Xin0 from alternative i and calculate its sample mean as
0
1 ∑
X̄i (n0 ) =
Xil .
n0 l=1
n
For any i, j with i ̸= j, calculate the sample variance of their difference,
2
Sij
(n0 )
n0
[
]2
1 ∑
Xir − Xjr − (X̄i (n0 ) − X̄j (n0 )) .
=
n0 − 1 r=1
Set n = n0 .
Update. If n > n0 , calculate the sample mean for each alternative i ∈ I, and for any pair of
2
alternatives i, j ∈ I with i ̸= j, update the sample variance Sij
(n) of their difference.
Screening. Let I old = I and
2
τij (n) = n/Sij
(n),
Define
and
Zij (τij (n)) = τij (n)[X̄i (n) − X̄j (n)].
{
}
I = I old \ i ∈ I old : Zij (τij (n)) ≤ −gc (τij (n)) for some other j ∈ I old .
Stopping Rule. If |I | = 1, stop and select the alternative whose index is in I as the best. Otherwise,
take one additional observation Xi,n+1 from each alternative i for i ∈ I, set n = n + 1 and go to
Update.
Remark 3. Similar to Procedure 1, Procedure 2 can also work with many choices of boundaries
√
(i.e., G ). As an instance, Procedure 2 uses a specific boundary gc (t) = [c + log(t + 1)](t + 1) and
the following results for Procedure 2 are shown for this specific boundary. Nevertheless, these
results can be easily extended to other boundaries in G .
Procedure 2 differs from Procedure 1 in two aspects. First, Procedure 2 determines the value of
c using an asymptotic result while Procedure 1 does it by solving a stochastic root-finding problem
(i.e., Equation (3)). Second, Procedure 2 updates the sample variances sequentially as more and
more observations are generated. This enables us to establish the asymptotic validity of Procedure
2 based on the strong consistency of the variance estimators.
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
15
4.2. Statistical Validity
Procedure 2 is only parameterized by the PCS value 1 − α. Therefore, we consider an asymptotic
regime, that is, PCS goes to 1 (or α → 0). Under this asymptotic regime, a procedure is asymptotically valid if its actual probability of incorrect selection (PICS) not only converges to 0 as α
converges to 0 but also converges at least as fast as α does, i.e., lim sup P{ICS}/α ≤ 1.
α→0
To show the asymptotic validity of Procedure 2, we need Lemma 6. It shows that the first-exit
time goes to infinity as α → 0 (or equivalently, c → ∞) with probability 1 (w.p.1). Based on this
lemma, we can prove that the variance estimators at the exit time converge to the true values as
α → 0. This allows us to treat the unknown variances essentially as their true values when analyzing
the asymptotic statistical validity of Procedure 2.
Lemma 6. Let {Yi : i = 1, 2, . . .} denote i.i.d. random variables with E[Y1 ] = µ, and define Z(n) =
∑n
+
i=1 Yi , for n = 1, 2, . . .. Let Nc = min{n ∈ Z : |Z(n)| ≥ gc (n)}, where gc (t) is defined in Procedure
2. Then Nc → ∞, w.p.1, as c → ∞.
Now we establish the asymptotic validity of Procedure 2 and summarize it in the following
theorem. The proof of this theorem is included in Appendix A.
Theorem 2. Suppose that there are k alternatives and we are interested in selecting the alternative
with the largest mean. Assume that (X1r , X2r , . . . , Xkr ), r = 1, 2, . . ., are independent and identically
jointly distributed with unknown mean vector (µ1 , µ2 , . . . , µk ), and their moment generating function
exists in a neighborhood of (0, 0, . . . , 0) ∈ Rk . Further, let the first-stage sample size n0 be a function
of α that satisfies n0 → ∞ as α → 0. If µ1 > µ2 > . . . > µk , Procedure 2 stops in finite time w.p.1
and satisfies lim sup P{alternative 1 is eliminated }/α ≤ 1.
α→0
Remark 4. To ensure the strong consistency of sequentially updated variance estimators, in addition to Lemma 6, we put some technical conditions on the first-stage sample size n0 , as listed
in Theorem 2. Although the theoretical condition facilitates the asymptotic proof, it does not
prescribe a specific setting of n0 in practice.
Similar to Proposition 1, we establish Proposition 2 for Procedure 2 in the case when the best
alternative is unique but the means are not all unique.
Proposition 2. Under the same assumptions as in Theorem 2, if µ1 > µ2 ≥ . . . ≥ µk , Procedure
2 terminates and eliminates alternative 1 with probability at most α in an asymptotic regime, i.e.,
lim sup P{terminate and eliminate alternative 1}/α ≤ 1.
α→0
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
16
5. Asymptotic Efficiency
When choosing a frequentist selection-of-the-best procedure, one not only requires the procedure to
deliver the guaranteed PCS, but also wants the procedure to be efficient, i.e., using as few observations as possible to select the best. To understand the efficiency of our procedures, we consider the
the general case when the observations are jointly generally distributed. In this situation, we focus
on our asymptotic procedure (i.e., Procedure 2) and study its average sample size asymptotically.
In Procedure 2, the strong consistency of variance estimators leads to an asymptotic analysis
that is similar to the known-variances case. Therefore, for simplicity, we assume that the variances
are known in this section. In Section 5.1 we provide an analytic expression for the asymptotic
average sample size required by Procedure 2, and in Section 5.2 we compare it to those of existing
IZ procedures.
5.1. Asymptotic Average Sample Sizes
As a building block, we consider a pairwise comparison between alternatives 1 and i with mean
difference µ1 − µi . Lemma 6 implies that the sample size required to detect them grows to infinity
in the asymptotic regime. In this case, their partial-sum difference process behaves more and more
like a line with slope µ1 − µi . Therefore, a heuristic method to calculate the average sample size
is to find the intersection point between the boundary and the line (see Figure 2). The method
is called the mean path approximation. Perng (1969) provides a rigorous proof of this heuristic
method for Paulson’s procedure and the proof can be extended easily to sequential IZ procedures
with a triangular continuation region (such as the KN procedure). The result is summarized in
Theorem 3.
(µ1 − µi )t
g(t)
Ni
−g(t)
Figure 2
Heuristic Mean Path Approximation Method
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
17
Theorem 3 (Perng (1969)). Suppose that there are k alternatives and we are interested in
selecting the alternative with the largest mean. Assume that (X1r , X2r , . . . , Xkr ), r = 1, 2, . . ., are
independent and identically jointly distributed with unknown mean vector (µ1 , µ2 , . . . , µk ) and a
positive definite covariance matrix. Further, assume µ1 − δ ≥ µ2 ≥ . . . ≥ µk , where δ > 0 is the IZ
parameter. Let (−aij + λij t, aij − λij t) (0 < λij < δ, 0 ≤ t ≤ aij /λij ) be the triangular region used
for a pairwise comparison between alternatives i and j. Define Ni as the sample size required to
distinguish µ1 and µi , for i = 2, 3, . . . , k, and define 1 − α as the probability of correct selection.
Then,
α→0
E[Ni ] ∼
2
where σ1i
= Var[X11 − Xi1 ].
2
a1i σ1i
, for all i = 2, 3, . . . , k,
µ1 − µi + λ1i
In particular, the KN procedure often chooses λij = δ/2 and aij = − 1δ log(2α/(k − 1)). Based
on Theorem 3, it is straightforward to derive a theoretical expression for the asymptotic average
sample size of the KN procedure, i.e.,
α→0
EKN [Ni ] ∼
2
f 2 σ1i
, for i = 2, 3, . . . , k,
(µ1 − µi + δ/2)δ
(9)
where f 2 = − log(2α/(k − 1)).
However, a rigorous proof has not been given so far for the sequential procedures with a general
boundary, such as the one used in Procedure 2. In next theorem, we accomplish the task and show
this heuristic method is also valid for Procedure 2 under mild conditions. The proof is included in
the appendix.
Theorem 4. Suppose that there are k alternatives and we are interested in selecting the alternative
with the largest mean. Assume that (X1r , X2r , . . . , Xkr ), r = 1, 2, . . ., are independent and identically
jointly distributed with unknown mean vector (µ1 , µ2 , . . . , µk ), and their moment generating function
exists in the neighborhood of (0, 0, . . . , 0) ∈ Rk . Let (−gc (t), gc (t)) denote the continuation region
√
in Procedure 2, where gc (t) = [c + log(t + 1)](t + 1). Define Ni as the sample size required to
distinguish µ1 and µi , for i = 2, 3, . . . , k, and define 1 − α as the probability of correct selection. If
µ1 > µ2 > . . . > µk , then, in Procedure 2, we have
2
c σ1i
α→0
, for all i = 2, 3, . . . , k,
E[Ni ] ∼
(µ1 − µi )2
2
= Var[X11 − Xi1 ].
where σ1i
(10)
Remark 5. Although the theorem is proven when the boundary function is gc (t) =
√
[c + log(t + 1)](t + 1), the proof can be easily extended to other boundaries.
Theorem 4 provides a theoretical foundation to evaluate the asymptotic average sample size
using the mean path approximation. For the comparison between alternatives 1 and i, it shows
that the asymptotic average sample size is inversely proportional to (µ1 − µi )2 , i.e., the square of
their true mean difference.
18
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
5.2. Comparisons between Procedure 2 and IZ Procedures
In this subsection we compare the asymptotic average sample sizes of Procedure 2 to the KN++
procedure. There are two reasons why we choose the KN++ procedure as the benchmark. First, the
KN++ procedure is the most similar procedure to Procedure 2. Both allow unknown and unequal
variances and CRNs, and both break the comparisons into pairwise comparisons and approximate
the partial sums of differences by Brownian motions. Second, the KN++ procedure, as the most
efficient one in the family of the KN procedures, has been well studied and compared in the
literature (Branke et al. (2007), Wang and Kim (2013)) and implemented by commercial simulation
software. Notice that it is not our intention to argue that the KN++ procedure is the most efficient
procedure in the literature. Indeed, it is not. When there are a large number of alternatives with
equal unknown variance, the BIZ procedure of Frazier (2014) is more efficient because it avoids the
use of Bonferroni inequality to break the comparisons into pairs, which causes inefficiency for the
KN++ procedure and our procedures as well. Therefore, when the number of alternatives is large,
we also expect the BIZ procedure to outperform ours. However, the objective of this comparison
is to show that the avoidance of setting IZ parameters may give an advantage to our procedures.
Even though the comparison is done between the KN++ procedure and Procedure 2, we believe
that the insights obtained here hold robustly for all IZ procedures.
Since both the KN++ procedure and Procedure 2 decompose the selection-of-the-best problem
into pairwise comparisons, we compare their asymptotic average sample sizes for the case of two
alternatives (i.e., k = 2). Let ∆ denote the difference of their means and set the variance of their
difference to 1. It is easy to calculate the asymptotic average sample size required by Procedure
2 based on Theorem 4. Meanwhile, given any specified IZ parameter δ, the asymptotic average
sample size of the KN++ procedure can be also evaluated according to Theorem 3.
To compare the asymptotic average sample sizes, we vary ∆/δ, which indicates how the chosen
IZ parameter relates to the true mean difference. In Figure 3, we plot the ratios of sample sizes of
√
the KN++ procedure and Procedure 2 with boundary [c + log(t + 1)](t + 1) for different values of
∆/δ. From Figure 3, we find that the KN++ Procedure needs about one fifth samples of Procedure
2 when ∆/δ = 1. However, when ∆/δ is above 5, Procedure 2 begins to out-perform the KN++
procedure. Particularly, as ∆/δ becomes larger and larger than 5, Procedure 2 requires fewer and
fewer samples than the KN++ procedure.
Even though Procedure 2 appears to require more samples when ∆/δ is smaller than 5, the
disadvantage may vanish when the number of alternatives is large. In this situation, there are often
many alternatives whose means are significantly smaller than the best. This may result in large
values of ∆/δ for many pairwise comparisons and thus increase the total number of observations
required by the KN++ procedure relative to Procedure 2.
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
Figure 3
Table 1
19
Ratios of asymptotic average sample sizes using the KN++ procedure and Procedure 2 (with the bound√
ary gc (t) = [c + log(t + 1)](t + 1) ) when ∆/δ varies.
Asymptotic average sample sizes of the Rinott’s procedure, the KN++ procedure and Procedure 2 to select from
two random variables whose difference is of mean ∆ and variance σ 2 .
Procedures
Asymptotic Sample Size
Rinott’s Procedure
KN++ Procedure
Procedure 2
h2 σ 2
δ2
f 2 σ2
(∆−δ/2)δ
cσ 2
∆2
Note: h2 is the Rinott’s constant (see Rinott 1978).
To close this section, we summarize the asymptotic average samples sizes required by Procedure
2, the KN++ procedure as well as the Rinott’s procedure, and list them in Table 1. For simplicity,
Table 1 considers a pairwise comparison where ∆ and σ 2 are the mean and variance of their
difference. We have two findings from the table and they are consistent with the findings from
Figure 3. First, both the KN++ and Rinott’s procedures are sensitive to the choice of δ. In particular,
the asymptotic sample sizes of the KN++ and Rinott’s procedures are inversely proportional to δ
and its square. Second, the asymptotic sample size of Procedure 2 is immune to the choice of δ
because Procedure 2 is IZ-free. Therefore, Procedure 2 is relatively most efficient if δ is chosen too
conservatively.
6. A Stopping Criterion
One drawback of Procedures 1 and 2 is that the continuation regions are open on the right-hand
side, which may potentially cause the procedures to run very long time before they stop or maybe
never stop. As a remedy, part (1) of Lemma 2 shows that these procedures can stop in finite time
w.p.1 in each pairwise comparison as long as the two alternatives have different means.
20
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
However, there might be two or more alternatives that have the same means. Then, there is
a positive probability that our procedures may not terminate (seen from Propositions 1 and 2 ),
resulting in an infinite sample size. To avoid this problem, we design truncated procedures, which
force the original procedures to stop when a stopping criterion is satisfied, even if there is more
than one alternative in contention.
Beyond avoiding an infinite sample size, users may have practical reasons for designing such
stopping criteria. For instance, they may have a practically significant difference below which difference in mean performances is not meaningful. This could arise if the resolution of the simulation
model itself is limited (e.g., nearest $10, 000), or if the precision of the software itself is finite (e.g.,
5 decimal digits is the default in Matlab). We refer to this limit as the error tolerance ε. Therefore,
the revised goal is to select an alternative that is within ε of the best, and to design a stopping
criterion that terminates once this goal is achieved.
Now we illustrate how to design a stopping criterion that corresponds to the error tolerance ε.
Denote Iε as the set of alternatives whose means are within ε of the best and the goal is to select
an alternative that is in Iε . Intuitively, we should choose a termination time T ∗ (ε) such that only
alternatives in Iε can survive at T ∗ (ε). In particular, we design T ∗ (ε) as follows:
Stopping Criterion Determine a termination time T ∗ (ε) that is a root to T ε − gc (T ) = 0.
Basically, the choice of T ∗ (ε) above is driven by the mean path approximation (see, Theorem
4). This method shows that the asymptotic average time required to eliminate all the alternatives
outside Iε is bounded by the value which is a root of T ε − gc (T ) = 0. In light of this, we determine
T ∗ (ε) as stated in Stopping Criterion above.
Provided the stopping criterion, we design a truncated procedure, i.e. Procedure 3, that is the
same as Procedure 2 except adding a stopping rule as follows: if |I | > 1 and τij (n) ≥ T ∗ (ε) for all
i, j ∈ I with i ̸= j, let i∗ = arg maxi∈I X̄i (n) and return i∗ . In the following theorem, we show
that Procedure 3 achieves the goal of selecting an alternative within ε of the best, with at least
the user-specified PCS in the limit.
Theorem 5. Assume that (X1r , X2r , . . . , Xkr ), r = 1, 2, . . ., are independent and identically jointly
distributed with unknown mean vector (µ1 , µ2 , . . . , µk ) and their moment generating function exists
in a neighborhood of (0, 0, . . . , 0) ∈ Rk . Further, let the first-stage sample size n0 be a function of
α that satisfies n0 → ∞ as α → 0. Without loss of the generality, assume that µ1 ≥ µ2 ≥ . . . ≥ µk .
Let ε denote the tolerance error and 1 − α denote the user-specified PCS. Let i∗ be the alternative
selected by Procedure 3. Then, we have, lim sup P{µi∗ < µ1 − ε}/α ≤ 1.
α→0
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
21
Remark 6. We refer to “correct selection” as selecting the best alternative and refer to “good
selection” as selecting any alternative within ε of the best. (The definitions are consistent with those
in Ni et al. (2014).) Theorem 5 states that Procedure 3 provides a good selection guarantee when
µ1 ≤ µ2 + ε. Furthermore, when µ1 > µ2 + ε, Procedure 3 provides a correct selection guarantee
because {µi∗ ≥ µ1 − ε} implies i∗ = 1. Therefore, Procedure 3 can provide both correct selection
and good selection guarantees.
If one treats ε as an IZ parameter, Procedure 3 turns into an IZ procedure because the selected
alternative is within ε of the best (shown by Theorem 5). Nevertheless, we want to emphasize that
the concept of an error tolerance is different from an IZ parameter. When the difference between
two alternatives is larger than ε, the average sample size required by the existing IZ procedures
depend critically on the choice of ε. In other words, a conservative ε may lead to excessive samples.
In contrast, a conservative ε will in general not affect the required sample size as our procedures
are expected to stop long before they reach the maximal sample size (due to Theorem 4). Another
way to see this is that the continuation regions of our procedures are not a function of ε. Therefore,
an error tolerance can be set very conservatively.
7. Numerical Experiments
In this section, we test the performance of our procedures through extensive numerical experiments
and compare them with existing procedures. Firstly, we establish the small-sample performance of
Procedures 1 and 2. Secondly, we consider the selection-of-the-best problem with various numbers of
alternatives. In this experiment, we compare the efficiency of Procedure 2 with the KN++ procedure
under three different configurations of variances. Thirdly, we test Procedure 3 that incorporates an
error tolerance and compare it with IZ procedures under different configurations of means. Lastly,
we test our procedures by addressing a realistic problem.
For simplicity, CRNs are not used in the experiments because using CRNs to increase their
efficiency is not our focus.
7.1. Small-Sample Performance
To get rid of the effect of using the Bonferroni inequality, we consider the case of only two alternatives. Assume that their observations are normally distributed with mean (µ1 , µ2 ) and variance
(5, 5). Further, assume alternative 1 is the better alternative which we would like to select. In this
experiment, we implement Procedures 1 and 2 to select alternative 1 with a desired PCS 0.95. For
different combinations of (µ1 , µ2 ) and the first-stage sample size n0 , we report the approximated
values of c, the estimated PCS and the average sample size with 95% confidence interval across
10,000 microreplications in Table 2 .
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
22
Table 2
The estimated PCS (Est. PCS) and the average sample sizes (Avg. SSize) with 95% confidence intervals
for Procedures 1 and 2 when observations are normally distributed.
µ1 − µ2
0.05
0.20
n0
10
15
20
25
30
35
40
10
15
20
25
30
35
40
Procedure 1
c
Est. PCS
Avg. SSize
7.358
0.97
(1.149 ± 0.019) × 105
5.870
0.97
(9.905 ± 0.160) × 104
5.163
0.97
(9.181 ± 0.148) × 104
5.040
0.97
(9.124 ± 0.142) × 104
4.671
0.97
(8.710 ± 0.138) × 104
4.514
0.97
(8.509 ± 0.136) × 104
4.415
0.98
(8.411 ± 0.133) × 104
7.358
0.99
(5.849 ± 0.097) × 103
5.870
0.99
(4.971 ± 0.083) × 103
5.163
0.99
(4.526 ± 0.078) × 103
5.040
0.99
(4.427 ± 0.074) × 103
4.671
0.99
(4.173 ± 0.070) × 103
4.514
0.99
(4.181 ± 0.070) × 103
4.415
0.99
(4.077 ± 0.068) × 103
Procedure 2
Est. PCS
Avg. SSize
0.96
(8.126 ± 0.125) × 104
0.97
(8.298 ± 0.127) × 104
0.97
(8.393 ± 0.127) × 104
4.605
0.97
(8.518 ± 0.126) × 104
0.98
(8.536 ± 0.126) × 104
0.98
(8.500 ± 0.125) × 104
0.98
(8.492 ± 0.125) × 104
0.98
(3.952 ± 0.067) × 103
0.99
(4.020 ± 0.065) × 103
0.99
(4.070 ± 0.067) × 103
4.605
0.99
(4.057 ± 0.066) × 103
0.99
(4.120 ± 0.066) × 103
0.99
(4.108 ± 0.067) × 103
0.99
(4.124 ± 0.060) × 103
c
From Table 2, we obtain several findings. First, both procedures deliver a larger PCS than the
desired PCS, and their delivered PCS decreases slightly as the true mean difference ∆ decreases
from 0.20 to 0.05. This arises because both procedures are designed for the case when the mean
difference is zero. Second, the average sample size of Procedure 1 depends on the setting of n0 .
In particular, for a given desired PCS, a smaller n0 leads to a larger sample size. This occurs
because Procedure 1 explicitly accounts for the (n0 − 1) degrees of freedom of the variance estimator
through using different values of c. Third, Procedure 1 often needs more samples than Procedure
2 does, but their difference tends to be smaller as n0 increases. Specially, as n0 increases to 40,
Procedure 1 needs fewer observations than Procedure 2 does. To understand this, as n0 increases,
the randomness coming from variance estimators fades away and the corresponding value of c for
Procedure 1 tends to be smaller and even less conservative (or smaller) than that of Procedure 2,
which is obtained by asymptotic approximation.
Table 2 shows that both procedures can be used to select the best in this experiment. Therefore,
users may face the choice between Procedures 1 and 2, and our suggestions are as follows. When
observations are normally distributed, Procedure 1 would be preferred unless it is significantly
less efficient than Procedure 2, because Procedure 1 guarantees PCS in small samples. Otherwise,
Procedure 2 would be preferred because it guarantees PCS although in an asymptotic regime while
Procedure 1 can not. Besides, Procedure 2 is more easily used because of its closed-form expressions
for the design parameter c and achieves a higher efficiency because it updates variance estimators
sequentially.
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
23
7.2. Large-scale Selection of the Best
In practice, the means of alternatives are often spread out when the number of alternatives is
large. In light of this, we consider a monotone decreasing configuration of means (µ1 , µ2 , . . . , µk )
where µi = 1.5 − 0.5i, and an equal-variance configuration where σi2 = 10 for all i. Our target is to
select alternative 1 with a desired PCS 0.95. To show the performance of our procedure to solve
large-scale selections of the best, we vary the number of alternatives as k = 20, 50, 100, 500.
To solve this problem, we use Procedure 2 for simplicity of the simulation study because its
design parameter c is more easily determined. As a comparison, we implement the KN procedure
by choosing an IZ parameter δ. It is worthwhile emphasizing that δ is chosen to enable the implementation of the KN procedure and we do not refer it to the smallest difference users care to detect
in this experiment. Further, to make a fair comparison, we compare Procedure 2 with the KN++
procedure because both of them update variance estimators sequentially. In this experiment, we
set n0 to 10.
Table 3
KN++
k
The estimated PCS and average sample sizes with 95% confidence interval for Procedure 2 and the
procedure under equal configuration of variances when all observations are normally distributed.
Procedure 2
0.99
20
2.816 × 103
±0.082 × 103
1.00
50
3.588 × 103
±0.089 × 103
1.00
100
4.388 × 103
±0.089 × 103
0.99
500
9.138 × 103
±0.102 × 103
1/4
0.70
2.639 × 102
±0.020 × 102
0.72
5.904 × 102
±0.023 × 102
0.76
1.112 × 103
±0.003 × 103
0.79
5.164 × 103
±0.003 × 103
1/2
0.91
5.503 × 102
±0.064 × 102
0.94
9.619 × 102
±0.075 × 102
0.95
1.540 × 103
±0.008 × 103
0.97
5.721 × 103
±0.009 × 103
KN++ procedure: (µ1 − µ2 )/δ
1
2
1.00
1.00
1.371 × 103
3.247 × 103
3
±0.016 × 10 ±0.029 × 103
0.99
1.00
2.014 × 103
4.454 × 103
±0.018 × 103 ±0.033 × 103
0.99
1.00
2.710 × 103
5.506 × 103
±0.020 × 103 ±0.034 × 103
1.00
1.00
7.217 × 103
1.077 × 104
±0.021 × 103 ±0.004 × 104
4
1.00
7.045 × 103
±0.050 × 103
1.00
9.779 × 103
±0.054 × 103
1.00
1.179 × 104
±0.006 × 104
1.00
1.877 × 104
±0.006 × 104
8
1.00
1.470 × 104
±0.007 × 104
1.00
2.070 × 104
±0.008 × 104
1.00
2.514 × 104
±0.009 × 104
1.00
3.628 × 104
±0.009 × 104
To compare Procedure 2 with the KN++ procedure, we consider different settings of δ where
(µ1 − µ2 )/δ = 1/4, 1/2, 1, 2, 4, 8. In Table 3, we report the estimated PCS and the average sample
sizes with 95% confidence intervals based on 1000 macroreplications of each procedure. From the
table, we obtain the following three findings. Firstly, when users select δ that is larger than µ1 − µ2
in the KN++ procedure, they may suffer from a lower PCS than they require, seen from the third
and fourth columns. Secondly, for any fixed number of alternatives (see each row in Table 3),
Procedure 2 demands fewer observations than the KN++ procedure, when δ is much smaller than
the true difference µ1 − µ2 . Furthermore, the gap of their sample sizes becomes larger as δ becomes
smaller. Thirdly, from the second and fifth columns, we can conclude that, although Procedure 2
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
24
needs more observations than the KN++ procedure when δ is exactly µ1 − µ2 , the relative difference
between their average sample sizes tends to be narrower as the number of alternatives k increases.
This implies the average sample size required by Procedure 2 grows more slowly with k than the
KN++ procedure.
Furthermore, Procedure 2 is shown to be asymptotically valid when the observations are generally
distributed. To show its small-sample performance for this general case, we repeat the experiment
above using the same parameters except that all the observations are exponentially distributed.
The results are reported in Tables 4 and they are similar to that of the normally distributed case.
Table 4
The estimated PCS and the average sample sizes with 95% confidence intervals of Procedure 2 and the
KN++ procedure when all observations are exponentially distributed.
k
Procedure 2
0.99
20
2.902 × 103
±0.084 × 103
0.99
50
3.610 × 103
±0.088 × 103
1.00
100
4.396 × 103
±0.092 × 103
1.00
500
9.095 × 103
±0.104 × 103
1/4
0.71
2.740 × 102
±0.022 × 102
0.74
5.972 × 102
±0.026 × 102
0.74
1.119 × 103
±0.003 × 103
0.79
5.167 × 103
±0.003 × 103
1/2
0.90
5.470 × 102
±0.067 × 102
0.95
9.430 × 102
±0.074 × 102
0.95
1.525 × 103
±0.008 × 103
1.00
5.699 × 103
±0.010 × 103
KN++ procedure: (µ1 − µ2 )/δ
1
2
0.99
1.00
1.344 × 103
3.143 × 103
3
±0.015 × 10 ±0.028 × 103
1.00
1.00
1.932 × 103
4.215 × 103
±0.017 × 103 ±0.030 × 103
1.00
1.00
2.627 × 103
5.306 × 103
±0.018 × 103 ±0.031 × 103
1.00
1.00
7.114 × 103
1.051 × 104
±0.018 × 103 ±0.004 × 104
4
1.00
6.927 × 103
±0.046 × 103
1.00
9.316 × 103
±0.052 × 103
1.00
1.124 × 104
±0.006 × 104
1.00
1.806 × 104
±0.006 × 104
8
1.00
1.463 × 104
±0.007 × 104
1.00
2.020 × 104
±0.008 × 104
1.00
2.414 × 104
±0.009 × 104
1.00
3.483 × 104
±0.009 × 104
Last, we further investigate the efficiency of Procedure 2 and the KN++ Procedure in equal,
increasing and decreasing configurations of variances where σi2 = 10, 10 × (0.95 + 0.05i) and
10/(0.95 + 0.05i) for i = 1, 2, . . . , k, respectively. To avoid reporting the similar results, in the following Table 5 we only the report the results when observations are normally distributed. From
this table, we find that the same conclusions seen from Tables 3 and 4 are valid.
7.3. Comparisons with IZ Procedures
In this subsection, we consider a selection-of-the-best problem with an error tolerance ε and solve
this problem using Procedure 3. If one treats ε as an IZ parameter, IZ procedures can also be used
to solve this problem. In this experiment we choose the (two-stage) Rinott’s procedure and the
(fully sequential) KN++ procedures as representatives of IZ procedures, and compare Procedure 3
with them in terms of the efficiency.
We consider two different configurations of means. The first one is the slippage configuration (SC)
where µ1 = 0.5, µ2 = µ3 = . . . = µ50 = 0. The second one is the monotone decreasing configuration
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
Table 5
25
The estimated PCS and average sample sizes with 95% confidence interval for Procedure 2 and the
KN++ procedure under different configurations of variances when all observations are normally distributed.
k
Variances
Decreasing
20
Equal
Increasing
Decreasing
50
Equal
Increasing
Decreasing
100
Equal
Increasing
Procedure 2
0.99
2.780 × 103
±0.081 × 103
0.99
2.816 × 103
±0.082 × 103
0.99
2.916 × 103
±0.085 × 103
0.99
3.551 × 103
±0.086 × 103
1.00
3.588 × 103
±0.089 × 103
0.99
3.635 × 103
±0.090 × 103
1.00
4.351 × 103
±0.093 × 103
1.00
4.388 × 103
±0.089 × 103
0.99
4.444 × 103
±0.093 × 103
1
0.99
1.326 × 103
±0.016 × 103
1.00
1.371 × 103
±0.016 × 103
0.99
1.473 × 103
±0.017 × 103
1.00
1.881 × 103
±0.017 × 103
0.99
2.014 × 103
±0.018 × 103
0.99
2.147 × 103
±0.018 × 103
1.00
2.592 × 103
±0.019 × 103
0.99
2.710 × 103
±0.020 × 103
0.99
2.934 × 103
±0.020 × 103
KN++ procedure: (µ1 − µ2 )/δ
2
4
1.00
1.00
3.092 × 103
6.720 × 103
3
±0.028 × 10 ±0.049 × 103
1.00
1.00
3.247 × 103
7.045 × 103
±0.029 × 103 ±0.050 × 103
1.00
1.00
3.423 × 103
7.437 × 103
±0.031 × 103 ±0.049 × 103
1.00
1.00
4.120 × 103
8.970 × 103
±0.033 × 103 ±0.052 × 103
1.00
1.00
4.454 × 103
9.779 × 103
±0.033 × 103 ±0.054 × 103
1.00
1.00
4.958 × 103
1.085 × 104
±0.034 × 103 ±0.005 × 104
1.00
1.00
5.061 × 103
1.060 × 104
±0.033 × 103 ±0.006 × 104
1.00
1.00
5.506 × 103
1.179 × 104
±0.034 × 103 ±0.006 × 104
1.00
1.00
6.350 × 103
1.412 × 104
±0.036 × 103 ±0.006 × 104
8
1.00
1.411 × 104
±0.007 × 104
1.00
1.470 × 104
±0.007 × 104
1.00
1.536 × 104
±0.008 × 104
1.00
1.910 × 104
±0.008 × 104
1.00
2.070 × 104
±0.008 × 104
1.00
2.268 × 104
±0.002 × 104
1.00
2.238 × 104
±0.009 × 104
1.00
2.514 × 104
±0.009 × 104
1.00
3.000 × 104
±0.009 × 104
of means (MDM) where µi = 1 − 0.5i for i = 1, 2, . . . , 50. For both configurations of means, we use
the equal-variance configuration where σi2 = 10, for all i = 1, 2, . . . , 50. For each configuration, we
select the best using three different procedures and report their average sample sizes based on 1000
macroreplications of each procedure in Table 6. We set PCS to 0.95 and n0 to 10.
From Table 6, we obtain the following insights. Firstly, the last column shows that the sample size
required by Procedure 3 appears stable as the ratio (µ1 − µ2 )/ε varies under MDM. It follows from
that, under MDM, Procedure 3 tends to terminate early before reaching the stopping criterion.
Secondly, Procedure 3 has a wider confidence interval under SC than the other two procedures.
This occurs because Procedure 3 either stops long before reaching the stopping criterion or stops
when the stopping criterion is met. This implies that Procedure 3 effectively avoids infinite samples.
Thirdly, from the first two columns, we find that the sample size required by the Rinott’s procedure
increases nearly 4 times as (µ1 − µ2 )/ε increases from 1 to 2, and increases nearly 25 times as
(µ1 − µ2 )/ε increases from 1 to 5. Similar results hold for the KN++ procedure although it increases
more slowly. Therefore, the Rinott’s procedure is the most sensitive to the choice of ε, then the
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
26
Table 6
Average sample sizes with 95% confidence intervals for the Rinott’s procedure, the KN++ procedure
and Procedure 3.
(µ1 − µ2 )/ε
1
2
5
(µ1 − µ2 )/ε
1
2
5
Normally Distributed Observations
SC
MDM
Rinott
KN++ Procedure 3
Rinott
KN++
6.071 × 104
1.576 × 104
4.781 × 104
6.045 × 104
2.002 × 103
±0.035 × 104 ±0.061 × 104 ±0.025 × 104
±0.025 × 104 ±0.018 × 103
5
4
4
2.421 × 10
4.022 × 10
6.430 × 10
2.426 × 105
4.425 × 103
5
4
4
5
±0.010 × 10 ±0.042 × 10 ±0.217 × 10
±0.010 × 10 ±0.033 × 103
6
5
4
1.515 × 10
1.129 × 10
8.857 × 10
1.518 × 106
1.254 × 104
6
5
4
6
±0.006 × 10 ±0.008 × 10 ±1.354 × 10
±0.006 × 10 ±0.006 × 104
Exponentially Distributed Observations
SC
MDM
Rinott
KN++ Procedure 3
Rinott
KN++
4
4
4
4
6.071 × 10
1.602 × 10
4.879 × 10
6.046 × 10
1.933 × 103
4
4
4
4
±0.046 × 10 ±0.015 × 10 ±0.058 × 10
±0.048 × 10 ±0.017 × 103
5
4
4
2.440 × 10
4.020 × 10
6.351 × 10
2.426 × 105
4.223 × 103
±0.019 × 105 ±0.035 × 104 ±0.148 × 104
±0.019 × 105 ±0.030 × 103
1.518 × 106
1.131 × 105
7.685 × 104
1.516 × 106
1.194 × 104
6
5
4
6
±0.012 × 10 ±0.007 × 10 ±0.985 × 10
±0.012 × 10 ±0.006 × 104
Procedure 3
3.119 × 103
±0.053 × 103
3.633 × 103
±0.091 × 103
3.615 × 103
±0.086 × 103
Procedure 3
3.117 × 103
±0.050 × 103
3.583 × 103
±0.087 × 103
3.646 × 103
±0.089 × 103
KN++ procedure and Procedure 3 is the least sensitive to the choice of ε. The results in Tables 6
are consistent with the theoretical results listed in Table 1. Lastly, the similar results appear for
both the normally and exponentially distributed cases.
7.4. Ambulance Allocation in Emergency Medical Services System
An Emergency Medical Services (EMS) system is an important component of public safety that
attempts to allocate scarce resources as critical events occur. In the EMS system, one critical
resource allocation challenge is deploying several ambulances to best serve requested emergency
calls, for instance, keeping response time small. The response time for a call is the elapsed time
from when the call is received to when an ambulance arrives at the scene. Here we measure the
performance as the fraction of requested calls received that have response times of 8 minutes or
less. In this experiment, we consider an example similar to that in Ni et al. (2012).
Consider a city of size 15 miles by 15 miles. Suppose there are 4 ambulances with traveling speed
24 miles per hour, 9 ambulance bases and 2 hospitals in the city. Distances from point to point
are measured by the Manhattan metric. See Figure 4 for a map of this city. When a requested call
arrives, the EMS system will dispatch the call to the nearest free ambulance. After the dispatched
ambulance arrives at the scene, the time spent there is exponentially distributed with mean 12
minutes before carrying the patient to the nearest hospital. Once the patient reaches the hospital,
the ambulance travels back to its home base to wait for the next call. We assume the call location
is uniformly distributed in the city. Our target is allocating these 4 ambulances to possible bases
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
27
to maximize the performance, i.e., the portion of calls with response time no more than 8 minutes.
In this experiment, we assume that more than one ambulance can be allocated to the same base.
15
miles
12
9
6
3
3
6
9
12
15
miles
Figure 4
Map for the city. (∗ stands for a possible ambulance base and ◦ stands for a hospital.)
To select the optimal policy in this example, we are faced with at least three challenges. First,
there are 495 alternative policies in this problem, which is not a small number for the classic
selection procedures. Second, the observations from each alternative policy are generally distributed
and with unknown variances. Last and the most important, the ideal case is to set the IZ parameter
to zero and select the best policy in this problem because when human life is concerned no difference
is so small that we are indifferent. In other words, when the IZ parameter is set as 0.01, it implies
that 0.01 × 100% of the citizens are in a dangerous situation because it may be fatal if their calls
are not handled in time. To circumvents these challenges, we use Procedure 2 that is IZ-free. In
our experiment, we set the PCS to 0.99 and replicate the selection process 1000 times. The result
shows that the optimal policy is selected as the best with probability 1.000 and the average total
sample size required is 2.10 × 104 , seen from Table 7.
One may argue that the IZ procedures (e.g. the KN++ procedure) can also be utilized to select the
optimal policy by choosing a conservative IZ parameter. To compare the performances of Procedure
3 and the KN++ procedure, we run the following experiments when ε is set as 0.01 and 0.001
respectively, and the results are listed in Table 7. Via Monte Carlo simulation, we evaluated the
performance of Policy 1, 2, 3, 4 as 0.8189, 0.8162, 0.8151 and 0.8144 because they are all selected
as the best in our experiments, and Policy 1 is the optimal policy.
From Table 7, we obtain the following conclusions. First, both procedures select a policy within
δ of the best with a probability greater than 0.99. Second, when δ is 0.01, Procedure 3 achieves
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
28
a larger PCS, although it costs about 1.5 times more samples than the KN++ procedure. As δ
decreases to 0.001, the sample size required by the KN++ procedure increases faster than that of
Procedure 3, implying that the KN++ procedure is more sensitive to the setting of δ.
Comparisons of our procedures and the KN++ procedure under different settings of the IZ parameter.
Table 7
δ
Good Selection
0.01
Policy
Policy
Policy
Policy
1
2
3
4
0.001
Policy 1
0
Policy 1
Procedure 3
Selection Prob. Avg. SSize
1.00
0.00
1.859 × 104
0.00
±0.013 × 104
0.00
2.102 × 104
1.00
±0.022 × 104
2.097 × 104
1.00
±0.021 × 104
KN++ Procedure
Selection Prob. Avg. SSize
0.93
0.04
7.943 × 103
0.02
±0.027 × 103
0.01
1.107 × 105
1.00
±0.003 × 105
NA
NA
8. Conclusion
In this paper, we propose a new frequentist selection-of-the-best formulation which selects the best
alternative with a user-specified probability of correct selection whenever the alternatives have
unique means. Under this formulation, we design a class of sequential procedures based on the Law
of the Iterated Logarithm. Among these procedures, the design parameters are obtained by solving
numerically a one-dimensional root-finding problem or using asymptotic approximating results.
We call them Procedures 1 and 2, and they are shown to be statistically valid in a finite regime
and an asymptotic regime, respectively. The merit of Procedures 1 and 2 is that they free users
from having to specify an IZ parameter. In addition, we add a stopping criterion to turn them into
an IZ procedure (Procedure 3). The numerical results show that our procedures are less sensitive
to the setting of IZ parameter and more efficient than the KN procedure when there are a large
number of alternatives with many significantly inferior ones.
Acknowledgement
The authors would like to thank Professor Shane G. Henderson from Cornell University for his
helpful discussions. A preliminary version of this paper (Fan and Hong 2014) was published in the
Proceedings of the 2014 Winter Simulation Conference.
Appendix A: Proofs of Lemmas and Theorems
Proof of Lemma 1
Proof. For any n ∈ Z + , 1 ∈ I(n ∧ N ) is equivalent to the event where either alternative 1 is
selected as the best eventually or it is eliminated after n, i.e.,
P{1 ∈ I(n ∧ N )} = P{select alternative 1} + P{eliminate alternative 1 after n}
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
29
≥ P{select alternative 1}
≥ 1 − α.
The second inequality holds due to Equation (1). Therefore, the conclusion is drawn.
2
Proof of Lemma 2
Proof. For statement (1), it is straightforward to deduce that Tc ≤ Td and thus it suffices to prove
that Td < ∞ with probability 1 (w.p.1). Define T̃d = min{t ∈ {ti : i = 1, 2, . . .} : B∆ (t) ≥ g(t)}. One
can easily check that T̃d ≥ Td and thus only need to prove T̃d < ∞ w.p.1. Letting B(t) = B∆ (t) − ∆t,
we know that lim B(ti )/ti = 0, w.p.1 (e.g., Karatzas and Shreve (1991), 2.9.3). And, g(t) ∈ G
i→∞
implies that g(t) = o(t). Therefore, we have
B∆ (ti ) − g(ti ) B(ti )
g(ti )
=
+∆−
→ ∆,
ti
ti
ti
a.s.
which implies lim (B∆ (ti ) − g(ti )) = ∞, w.p.1, and hence T̃d < ∞, w.p.1.
i→∞
Given statement (1), statement (2) has been shown in Jennison et al. (1980).
2
Proof of Lemma 3
Proof. Denote Z(t, ω) as a sample path from the stochastic process Z(t) and let Z∆ (t, ω) =
Z(t, ω) + ∆t. Further, let T (ω) = inf {t : |Z(t, ω)| ≥ g(t)} and T∆ (ω) = inf {t : |Z∆ (t, ω)| ≥ g(t)}. To
justify this lemma, it suffices to show that,
{ω : Z∆ (T∆ (ω), ω) ≤ −g(T∆ (ω)), T∆ (ω) < ∞} ⊆ {ω : Z(T (ω), ω) ≤ −g(T (ω)), T (ω) < ∞}. (11)
In the rest of this proof, we show (11) by contradiction. For any sample path ω, when
Z∆ (T∆ (ω), ω) ≤ −g(T∆ (ω)) holds with T∆ (w) < ∞, we readily obtain that
Z(T∆ (ω), ω) = Z∆ (T∆ (ω)) − T∆ (ω)∆ ≤ −g(T∆ (ω)),
which implies that Z(·, ω) exits the continuation region no later than T∆ (ω), i.e.,
T (ω) ≤ T∆ (ω) < ∞.
(12)
Suppose that {Z(T (ω), ω) ≤ −g(T (ω)), T (ω) < ∞} is not true, i.e., Z(T (ω), w) ≥ g(T (ω)) with
T (ω) < ∞. Then we can derive
Z∆ (T (ω), ω) = Z(T (ω), ω) + T (ω)∆ ≥ g(T (ω))
(13)
which implies that Z∆ (·, ω) exits from the continuation region no later than T (ω), i.e.,
T∆ (ω) ≤ T (ω) < ∞.
(14)
From inequalities (12) and (14), we have T∆ (ω) = T (ω) < ∞. Plugging T∆ (ω) = T (ω) into inequality
(13) yields Z∆ (T∆ (ω), ω) ≥ g(T∆ (ω)), which contradicts from Z∆ (T∆ (ω), ω) ≤ −g(T∆ (ω)). Therefore, this lemma is drawn.
2
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
30
Proof of Lemma 6
Proof. Denote Z(·, ω) as a sample path from the stochastic process Z(·) and Nc (ω) = min{n ∈ Z + :
|Z(n, ω)| ≥ gc (n)}. We show this lemma by contradiction. Let A = {ω : limc→∞ Nc (ω) ̸= ∞}. If the
conclusion is not true, then P{A} > 0. Notice that, for any sample path ω, Nc (ω) is non-decreasing
in c since gc (·) is increasing in c, and Nc (ω) only takes integer values. Therefore, A = ∪∞
n=1 An with
An = {ω : limc→∞ Nc (ω) = n}. Since P{A} > 0, there exists some constant integer N ∗ such that
P{AN ∗ } > 0. Further, for any ω ∈ AN ∗ , there exists a constant c∗ (ω) such that Nc (w) ≡ N ∗ for any
c ≥ c∗ (ω). Therefore, for any ω ∈ AN ∗ and c ≥ c∗ (ω),
|Z(N ∗ , ω)| = |Z(Nc (ω), ω)| ≥ gc (Nc (ω)) = gc (N ∗ ).
Letting c → ∞, we have |Z(N ∗ , ω)| = ∞ for any ω ∈ AN ∗ . Therefore, P{ω : |Z(N ∗ , ω)| = ∞} ≥
P{AN ∗ } > 0, contradicting with E[Z(N ∗ )] = N ∗ µ < ∞.
2
Proof of Theorem 2
Proof. For any pair of alternatives i and j with µi ̸= µj , we can follow the same logic (i.e. using
the Law of Iterated Logarithm) for proving part (1) of Lemma 2 and further show the elimination
between alternatives i and j occurs in finite time w.p.1. Then, noticing that the alternatives have
unique means, we conclude that all the eliminations occur in finite time w.p.1, implying that
Procedure 2 stops in finite time w.p.1.
Given the procedure stops in finite time, we define an incorrect-selection event happens when
alternative 1 is eliminated by some other alternative before termination of the procedure. Hence,
we can study the probability of each pairwise incorrect-selection (ICS) event separately and bound
the total probability of ICS using the Bonferroni inequality, i.e.,
P{ICS} ≤
k
∑
P{ICSi },
(15)
i=2
where ICSi denotes the incorrect-selection event that alternative 1 is eliminated by alternative i,
for i = 2, 3, . . . , k. To show the conclusion, we only need to show that
lim sup P{ICSi }/(α/(k − 1)) ≤ 1, for all i ̸= 1.
α→0
In particular, by Procedure 2, we have
P{ICSi } = P {Z1i (τ1i (Ni )) ≤ −gc (τ1i (Ni )), Ni < ∞} .
(16)
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
31
where Ni = min{n ≥ n0 : |Z1i (τ1i (n))| ≥ gc (τ1i (n))}. For the ease of theoretical analysis, throughout
this proof, we consider the case of single-sided boundary instead of the case of double-sided boundary and their connection can be built as follows. Let Ni− = min{n ≥ n0 : Z1i (τ1i (n)) ≤ −gc (τ1i (n))},
then we get that Ni− ≥ Ni and
{
(
)
}
P{ICSi } ≤ P Z1i (τ1i (Ni− )) ≤ −gc τ1i (Ni− ) , Ni− < ∞
(
)
}
{
= P Z1i (t1i (Ni− )) ≤ −gc t1i (Ni− )Y −1 (Ni− ) Y (Ni− ), Ni− < ∞
2
2
(Yi (n) = S1i
(n)/σ1i
).
(17)
First, we would like to show that in the asymptotic regime α → 0, (17) is upper bounded by the
corresponding probability that replaces the variance estimator with the true variance, or equivalently,
lim sup
α→0
(17)
≤ lim sup
α/(k − 1)
α→0
}
{
P Z1i (t1i (Ñi− )) ≤ −gc (t1i (Ñi− )), Ñi− < ∞
α/(k − 1)
,
(18)
where Ñi− = min{n ≥ n0 : Z1i (t1i (n)) ≤ −gc (t1i (n))}. This is ensured by the strong consistency of
variance estimators and the detailed clarification is as follows. Due to Lemma 6 and the fact that
Ni− ≥ Ni , we have Ni− → ∞ as α → 0. Further, since Y (n) → 1 w.p.1, for any small number ϵ, there
exists some n0 (< Ni− ) large enough, such that
P {|Y (n) − 1| ≤ ϵ for all n ≥ n0 } ≥ 1 − ϵ.
Also notice that the function h(t, y) := gc (t/y)y is an increasing function of y, as long as c > 1.
Actually, when c > 1,
∂
1
log h(t, y) = −
∂y
2
(
t
t
+
y(t + y)(c + log(t/y + 1)) y(t + y)
)
+
1
t
1
>−
+ > 0.
y
y(t + y) y
Therefore, we obtain
}
{
(1−ϵ)
(1−ϵ)
(1−ϵ)
(17) ≤ P Z1i (t1i (Ni
)) ≤ −h(t1i (Ni
), 1 − ϵ), Ni
< ∞; |Y (n) − 1| ≤ ϵ for all n ≥ n0
+P {|Y (n) − 1| > ϵ for some n ≥ n0 }
{
}
(1−ϵ)
(1−ϵ)
(1−ϵ)
≤ P Z1i (t1i (Ni
)) ≤ −h(t1i (Ni
), 1 − ϵ), Ni
< ∞ + ϵ,
(y)
where Ni
= min{n ≥ n0 : Z1i (t1i (n)) ≤ −h(t1i (n), y)}. Now set ϵ = o(α) and let α → 0, we obtain
{
}
(1)
(1)
(1)
P
Z
(t
(N
))
≤
−
h(t
(N
),
1),
N
<
∞
1i
1i
1i
i
i
i
(17)
lim sup
≤ lim sup
α/(k
− 1)
α/(k − 1)
α→0
α→0
{
}
P Z1i (t1i (Ñi− )) ≤ −gc (t1i (Ñi− )), Ñi− < ∞
= lim sup
.
(19)
α/(k − 1)
α→0
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
32
In other words, (18) holds.
−2
Furthermore, noticing that Z1i (t1i (n)) = σ1i
∑n
j=1 (X1j
− Xij ), we define a standardized version
of Z1i (t1i (·)) with a scaling parameter M (α), that is
∑⌊M (α)s⌋
−2
σ1i
[X1j − Xij − (µ1 − µi )]
j=1
√
C(M (α), t1i (s)) =
, for s ≥ 0,
M (α)
(20)
where we let M (α) → ∞ as α → 0. Based on Equation (8), we get that C(M (α), t1i (s)) → B(t1i (s)),
as M (α) → ∞. Plugging C(M (α), ·) into (18), we get that
{
}
(18) ≤ lim sup P Z1i (t1i (Ñi0− )) − t1i (Ñi0− )(µ1 − µi ) ≤ −gc (t1i (Ñi0− )), Ñi0− < ∞ /(α/(k − 1))
α→0
)
(
 (

(
))
0−


0−
)
g
t
(
Ñ
c
1i
i
Ñi
= lim sup P C M (α), t1i
≤− √
, Ñi0− < ∞ /(α/(k − 1))


M (α)
α→0
M (α)
}
{
(
) √
, lim sup P C(M (α), Td− ) ≤ −gc M (α)Td− / M (α), Td− < ∞ /(α/(k − 1))
α→0
= lim sup P{Td− < ∞}/(α/(k − 1)),
(21)
α→0
where Ñi0− = min{n ≥ n0 : Z1i (t1i (n)) − t1i (n)(µ1 − µi ) ≤ −gc (t1i (n)) } and Td− = t1i (Ñi0− /M (α)).
The inequality above follows from Lemma 3.
Next, we would like to show that (21) is upper bounded by the probability that considers a
continuous-time Brownian motion instead of C(M (α), ·) when
M (α) → ∞, c3 /M (α) → 0, as α → 0.
(22)
In the rest of the proof, for the simplicity of notation, we write M instead of M (α).
To show this result, we start with studying the asymptotic cumulative distribution function
P{Td− ≤ t} for any fixed t > 0. Because X11 − Xi1 has a finite moment generating function in the
neighborhood of zero, then by Theorem 2.2.4 in Whitt (2002b), there exist positive constants C1 ,
C2 , λ and a BM B̃(·) depending on the distribution of Xi1 and X11 , such that for every M ≥ 1,
{
}
(
)
C2 exp(−c)
−1
2
P max n[X̄1 (n) − X̄i (n) − (µ1 − µi )]/σ1i − B̃(n) ≥ C1 log s + λ
c − log M σ1i ≤
.
2
1≤n≤s
M σ1i
(23)
√
2
2
t)/ M σ1i
Define B(·) = B̃(M σ1i
. It is easy to see B(·) is still a BM by the self-similarity scaling
√
2
2
2
)]/ M σ1i
u + λ−1 (c − log M σ1i
property of BM. Let ψM (u) = [C1 log M σ1i
, then for any fixed t > 0,
P{Td− ≤ t}
=P{Td− ≤ t, |C(M, q) − B(q)| ≥ ψM (q), ∃ q = t1i (n/M ) ≤ t, n ∈ Z + }
+ P{Td− ≤ t, |C(M, q) − B(q)| ≤ ψM (q), ∀ q = t1i (n/M ) ≤ t, n ∈ Z + }
(24)
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
33
We bound the two probability terms in (24) separately. For the first term in (24), by the definition
of C(M, ·) in (20), we have
the first term in (24)
≤ P{|C(M, q) − B(q)| ≥ ψM (q), ∃ q = t1i (n/M ) ≤ t, n ∈ Z + }
2
⌊M σ1i
t⌋
≤
∑
P {|C(M, t1i (s/M )) − B(t1i (s/M ))| ≥ ψM (t1i (s/M ))}
s=1
{
}
n[X̄1 (n) − X̄i (n) − (µ1 − µi )]/σ1i
B̃(n)
√
√ ≥ ψM (t1i (s/M ))
P max ≤
−
1≤n≤s
σ
M
σ
1i
1i M
s=1
≤ C2 t exp(−c) by (23).
2
⌊M σ1i
t⌋
∑
(25)
To bound the second term in (24), we define a new single lower boundary −L(u), where L(u) =
√
gc (M u)/ M − ψM (u). Then, {Td− ≤ t, |C(M, q) − B(q)| ≤ ψM (q), ∀ q = t1i (n/M ) ≤ t, n ∈ Z + }
implies that B(·) crosses the single lower boundary −L(u) before t, that is
−
the second term of (24) ≤ P(Td,L
≤ t) ≤ P(TL− ≤ t),
(26)
−
where Td,L
= min{q = t1i (n/M ) : B(q) ≤ −L(q), n ∈ Z + } and TL− = inf {u : B(u) ≤ −L(u)} denote a
discrete and continuous stopping time for the one-sided lower boundary, respectively. The second
−
inequality arises from the fact that Td,L
≥ TL− . Actually, we can always find a large enough M that
satisfies the conditions in (22) to ensure a well-defined boundary L(u), i.e., L(u) > 0 for 0 < u ≤ t.
√
To see it, when M satisfies the conditions in (22), we have ψM (u) → 0 and gc (M u)/ M → ∞ as
α → 0, for any fixed u.
Combining (25) and (26), we obtain an upper bound for P{Td− ≤ t}, that is
P{Td− ≤ t} ≤ P(TL− ≤ t) + C2 t exp(−c).
(27)
Lastly, we calculate the asymptotic expression for P(TL− ≤ t) based on Theorem 4 in Jennen and
Lerche (1981). This theorem provides the asymptotic probability density of TL− , that is
(
)
′
L2 (u)
α→0 L(u) − uL (u)
√
fT − (u) ∼
exp −
, uniformly in (0, t) when α is sufficiently close to 0.
L
2u
2πu3/2
√
Let g̃c,M (u) = gc (M u)/ M and Tg̃−c,M = inf {t : B(t) ≤ −g̃c,M (t)}. Then we derive that
]
[
(
)
′
2
fT − (u) α→0
(u)
2g̃c,M (u)ψM (u) − ψM
ψM (u) − uψM (u)
L
exp
∼ 1+
.
′
fT − (u)
2u
g̃c,M (u) − ug̃c,M (u)
g̃c,M
Since c3 /M → 0 as α → 0, we get that g̃c,M (u)ψM (u) = o(1) as α → 0. Besides, by the definition
′
(u) → 0 as M → ∞ for any
of g̃c,M (u) and ψM (u), we know that g̃c,M (u) → ∞, ψM (u) → 0 and ψM
fixed u. Therefore,
α→0
fT − (u) ∼ fT −
L
g̃c,M
(u), uniformly on u ∈ (0, t].
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
34
from which we have
P{TL−
∫
t
α→0
< t} ∼
fT −
(u)du
√
√
( 2
)
∫ Mt
gc (M u) / M − ugc′ (M u) M
gc (M u)
α→0
√
∼
exp −
du
2M u
2πu3/2
0
( 2
)
∫ Mt
gc (v) − vgc′ (v)
gc (v)
√
=
exp −
dv
3/2
2v
2πv
0
[
(√
)]
α→0
∼ exp(−c/2) 1 − Φ
(c + log(M t + 1))/(M t) ,
0
g̃c,M
(28)
where Φ(·) denotes the cumulative distribution function of a standard normal random variable.
The second asymptotics arises from Equation (19) in Jennen and Lerche (1981). Plugging (28) into
(27) leads to
[
(√
)]
lim sup P{Td− ≤ t}/(α/(k − 1)) ≤ lim sup 2 1 − Φ
[c + log(M t + 1)]/(M t) .
α→0
(29)
α→0
Further, letting t → ∞ in (29) and combining it with (17), (18), (21), we have (16) holds and then
2
our conclusion is drawn.
Proof of Theorem 4
Proof. The proof contains two parts. Part one, we justify the mean path approximation, i.e.,
α→0
E[Ni ] ∼ λi , where λi denotes the intersection point between the boundary gc (t1i (λ)) and (µ1 −
µi )t1i (λ). Part two, we show λi can be further asymptotically rewritten as an explicit form, as
shown in the theorem.
To prove the first part, it suffices to prove the following two statements: (1) lim sup E[Ni ]/λi ≤ 1;
α→0
(2) lim inf E[Ni ]/λi ≥ 1, for all i = 2, 3, . . . , k. The proof can be summarized as follows.
α→0
To prove statement (1), we construct an upper linear support for the current boundary and show
the average sample size is λi asymptotically under this linear boundary. Then we claim that our
procedure has a smaller asymptotic average sample size than the one using the linear boundary.
As for statement (2), we show that it holds for a subset of the event space, thus it holds for the
whole event space.
We start with statement (1). Since gc (t) is concave, gc′ (t1i (λi )) t + t1i (λi )(µ1 − µi − gc′ (t1i (λi ))
builds an upper tangent line to gc (t) at the point (t1i (λi ), (µ1 − µi )t1i (λi )), i.e., gc (t) ≤ gc′ (t1i (λi ))t +
t1i (λi )(µ1 − µi − gc′ (t1i (λi ))) for any t. Let Ñi denote the first exit time of Z1i (t1i (·)) from the tangent
line, i.e.,
Ñi = min{n : Z1i (t1i (n)) ≥ gc′ (t1i (λi ))t1i (n) + t1i (λi )(µ1 − µi − gc′ (t1i (λi )))}.
It is easy to derive that
Ni = min{n : |Z1i (t1i (n))| ≥ gc (t1i (n))} ≤ Ñi .
(30)
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
35
One can check that the limit of gc′ (t1i (λi )) exists by noticing c, λi → ∞ as α → 0, and then due to
Lemma 2 in Siegmund (1967), we get that lim sup E[Ñi ]/λi ≤ 1. Further, by (30), we see that
α→0
lim sup E[Ni ]/λi ≤ lim sup E[Ñi ]/λi ≤ 1.
α→0
(31)
α→0
Now we move to justify the statement (2). Let (Ω, G , P) be the underlying probability space and
let Ω0 = {ω : alternative 1 eliminates alternative i}. By the strong law of large numbers, we have
X̄1 (n) − X̄i (n) converges to µ1 − µi almost surely on Ω0 . By Egoroff’s theorem (Egoroff 1911), for
every ϵ ∈ (0, µ1 − µi ) there exist disjoint subsets A and B, such that A ∪ B = Ω0 , P(B) < ϵ and
X̄1 (n) − X̄i (n) converges uniformly on A. Hence, there exists a positive integer n(ϵ), s.t. for every
ω ∈ A and n > n(ϵ), we have
0 < µ1 − µi − ϵ ≤ X̄1 (n) − X̄i (n) ≤ µ1 − µi + ϵ.
Define A∗ = {ω : ω ∈ A and Ni ≤ n(ϵ)}. Then, on the set A − A∗ , we have
0 < µ1 − µi − ϵ ≤ X̄1 (Ni ) − X̄i (Ni ) ≤ µ1 − µi + ϵ.
and hence
gc (t1i (Ni )) ≤ |t1i (Ni )[X̄1 (Ni ) − X̄i (Ni )]| = t1i (Ni )[X̄1 (Ni ) − X̄i (Ni )] ≤ t1i (Ni )(µ1 − µi + ϵ).
This implies that Ni ≥ λϵi on A − A∗ , where λϵi is such that gc (t1i (λϵi )) = t1i (λϵi )(µ1 − µi + ϵ).
Therefore, we can derive that
lim inf E[Ni ]/λi ≥ lim inf E[Ni I(w ∈ A − A∗ )]/λi ≥ lim P(A − A∗ )λϵi /λi .
α→0
α→0
α→0
According to Lemma 6, we know that Ni → ∞ as α → 0 w.p.1. Hence, we have lim P(A∗ ) = 0, and
α→0
lim P(A − A∗ ) = lim [P(A) − P(A∗ )] = P(A) = 1 − P(B) ≥ 1 − ϵ.
α→0
α→0
(32)
In other words, lim inf E[Ni ]/λi ≥ (1 − ϵ)λϵi /λi . Letting ϵ → 0, we have the second statement is true,
α→0
i.e., lim inf α→0 E[Ni ]/λi ≥ 1. Combining with Inequality (31), we get
α→0
E[Ni ] ∼ λi .
(33)
Now we move to the second part. Since λi satisfies the equation (µ1 − µi )t1i (λ) =
√
[c + log(t1i (λ) + 1)](t1i (λ) + 1), we can derive c = (µ1 − µi )2 t21i (λi )/(λi +1) − log(t1i (λi )+1). Then,
c
t1i (λi )
log(t1i (λi ) + 1)
=
−
.
2
t1i (λi )(µ1 − µi )
t1i (λi ) + 1 t1i (λi )(µ1 − µi )2
(34)
α→0
As α → 0, we have t1i (λ) → ∞. Consequently, (34) ∼ 1 and in other words,
2
c
cσ1i
α→0
=
∼ 1.
2
2
t1i (λi )(µ1 − µi )
λi (µ1 − µi )
Combining (33) and (35) results in the conclusion.
(35)
2
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
36
Proof of Theorem 5
Proof. If µ1 − µk ≤ ε, then the conclusion is trivially valid. Otherwise, there exists some integer
v such that µ1 − µv ≤ ε and µ1 − µv+1 > ε. To avoid the confusion, we define alternative i kills
alternative j(̸= i) if alternative i eliminates alternative j before T ∗ (ε) or alternative i has a larger
sample mean at T ∗ (ε), and define alternative i eliminates alternative j if the partial-sum process
of their mean difference exits the designed continuation region from above. (The definition of
“elimination” is consistent with the rest of this paper.) Then
P{µ1 − µi∗ ≤ ε}
≥ P {alt. v + 1, v + 2, . . . , k are killed}
{ v
}
k
∩
∩
≥P
{alt. 1 is not eliminated by alt. j before T ∗ (ε)}
{alt. 1 kills alt. i}
j=2
≥ 1−
v
∑
i=v+1
P {alt. 1 is eliminated by alt. j before T ∗ (ε)} −
j=2
k
∑
P {alt. i kills alt. 1} .
(36)
i=v+1
The last inequality arises from Bonferroni inequality. Throughout this proof, we define the
following notations. For all l ̸= 1, define two stopping times Nl and Nl0 as Nl = min{n :
0
0
Z1l (τ1l (n)) ≥ gc (τ1l (n))} and Nl0 = min{n : Z1l
(τ1l (n)) ≥ gc (τ1l (n))}, where Z1l
(τ1l (n)) =
Z1l (τ1l (n)) − τ1l (n)(µ1 − µl ). Then, for j = 2, 3, . . . , v, we have
P{alt. 1 is eliminated by alt. j before T ∗ (ε)}
= P{Z1j (τ1j (Nj )) ≤ −gc (τ1j (Nj )), τ1j (Nj ) ≤ T ∗ (ε) < ∞}
≤ P{Z1j (τ1j (Nj )) ≤ −gc (τ1j (Nj )), τ1j (Nj ) < ∞}
0
≤ P{Z1j
(τ1j (Nj0 )) ≤ −gc (τ1j (Nj0 )), τ1j (Nj0 ) < ∞}
α→0
∼ α/(k − 1).
(37)
The first inequality holds because the condition {τ1j (Nj ) ≤ T ∗ (ε)} is omitted and the second
inequality holds due to Lemma 3. The asymptotic equivalence is seen by the proof of Theorem 2.
For i = v + 1, . . . , k, alternative i kills alternative 1 if alternative 1 is eliminated by alternative i
before T ∗ (ε) or alternative 1 has a smaller sample mean at T ∗ (ε). In other words,
P{alt. i kills alt. 1}
=P{Z1i (τ1i (Ni )) ≤ −gc (τ1i (Ni )), τ1i (Ni ) ≤ T ∗ (ε) < ∞} + P{Z1i (T ∗ (ε)) < 0, T ∗ (ε) < τ1i (Ni )}. (38)
Now we study these two terms in (38) separately. For the first term, we have,
{ 0
}
the first term in (38) ≤ P Z1i
(τ1i (Ni0 )) ≤ −gc (τ1i (Ni0 )), τ1i (Ni0 ) < ∞, τ1i (Ni ) ≤ T ∗ (ε) < ∞ , (39)
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
37
which follows from Lemma 3. As for the second term in (38), we have that,
the second term in (38)
{ 0 ∗
}
= P Z1i
(T (ε)) ≤ −T ∗ (ε)(µ1 − µi ), T ∗ (ε) < τ1i (Ni )
{ 0 ∗
}
≤ P Z1i
(T (ε)) ≤ −ε T ∗ (ε), T ∗ (ε) < τ1i (Ni )
(since µ1 − µi > ε)
}
{ 0 ∗
(by the definition of T ∗ (ε)).
= P Z1i (T (ε)) ≤ −gc (T ∗ (ε)), T ∗ (ε) < τ1i (Ni )
(40)
0
In the probability statement above, the first statement implies that Z1i
(·) first exits the contin-
uation region no later than T ∗ (ε) (i.e., τ1i (Ni0 ) ≤ T ∗ (ε) < ∞) and the second statement further
0
0
claims that Z1i
(·) first exits the region from below. Otherwise, if Z1i
(τ1i (Ni0 )) ≥ gc (τ1i (Ni0 )), then
0
Z1i (τ1i (Ni0 )) = Z1i
(τ1i (Ni0 )) + τ1i (Ni0 )(µ1 − µi ) ≥ gc (τ1i (Ni0 )), implying that τ1i (Ni ) ≤ τ1i (Ni0 ) ≤
T ∗ (ε). This contradicts with T ∗ (ε) < τ1i (Ni ). Therefore, we validate this claim and then get that
the second term in (38) is upper bounded by
{ 0
}
P Z1i
(τ1i (Ni0 )) ≤ −gc (τ1i (Ni0 )), τ1i (Ni0 ) < ∞, T ∗ (ε) < τ1i (Ni ) .
Plugging (39) and (41) into (38), we obtain that
{ 0
} α→0
P{alt. i eliminates alt. 1} ≤ P Z1i
(τ1i (Ni0 )) ≤ −gc (τ1i (Ni0 )), τ1i (Ni0 ) < ∞ ∼
(41)
α
.
k−1
(42)
The asymptotic equivalence can be derived based on the proof of Theorem 2. Combining (36) with
2
(37) and (42), the conclusion is drawn.
Proof of Proposition 2
Proof. For any sample path, part (1) of Lemma 2 implies that Procedure 2 does not terminate in
finite time only if all the remaining alternatives have the unique mean. Because the best alternative
is unique, it suffices to state that the procedure does not terminate in finite time only if alternative
1 is eliminated. Further, we have
{
P{eliminate alternative 1} = P
k
∪
}
{alernative 1 is elimianted by alternative i}
i=2
≤
k
∑
P{alernative 1 is elimianted by alternative i}.
(43)
i=2
The inequality above arises from the Bonferroni inequality. Notice that (43) is the same as (15)
which is shown to be at most α asymptotically in the proof of Theorem 2. Therefore, we have
lim sup P{do not terminate in finite time}/α ≤ lim sup P{eliminate alternative 1}/α ≤ 1,
α→0
α→0
and
lim sup P{terminate and eliminate alterantive 1}/α ≤ 1.
α→0
Therefore, the proposition holds.
2
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
38
Appendix B: Simulation Based Algorithm for Finding c in Equation (3)
In this appendix, we design a simulation-based algorithm to find c(> 0) by solving Equation (3),
i.e.,
}]
)
[ {
(
Y
n0 − 1
α
T
, T < ∞Y
,
β , E P B(T ) ≤ −gc
=
Y
n0 − 1
k−1
{
}
(
)
where gc (t) ∈ G , T = inf t : |B(t)| ≥ gc n0Y−1 t n0Y−1 , B(·) denotes a Brownian motion with unit
variance and no drift, and Y is a chi-squared random variable with n0 − 1 degrees of freedom
that is independent of B(·). Solving (3) can be treated as a stochastic root-finding problem. To
make the problem well defined, it is worthwhile mentioning that Equation (3) is solvable. This
is guaranteed because the boundary function gc (t) ∈ G satisfies a certain lower bound condition
√
(i.e., it approaches to the infinity at a rate no slower than O( t log log t )). If this condition is not
satisfied, by the Law of the Iterated Logarithm, the left-hand side of Equation (3) is always 1/2
and thus (3) may have no solution.
(
{
)
}
To solve Equation (3), the key issue is to evaluate P B(T ) ≤ −gc n0y−1 T n0y−1 , T < ∞ for any
c > 0 when Y is observed as y. Once the probability has been evaluated, Equation (3) can be
addressed using linear interpolation, bisection method, Newton’s method and so on. Because gc (·)
can be a general boundary from G , it is difficult to obtain an explicit expression for this probability
and we have to estimate it via Monte Carlo simulation.
To estimate the probability, we generate a sample path from B(·) under the probability measure
P and check whether this sample path exits the boundary in finite time. However, this may be
infeasible because P{T = ∞} > 0. When {T = ∞} occurs for some sample path, the simulation
process fails to terminate in finite time. We overcome this difficulty by using a truncation method, in
(
{
)
}
which we estimate P B(T ) ≤ −gc n0y−1 T n0y−1 , T ≤ N (y) . The truncating point N (y) is chosen
to bound the truncation error below a pre-determined level η ∈ (0, 1), i.e.,
{
(
)
}
n0 − 1
y
T
, N (y) < T < ∞ ≤ η.
P B(T ) ≤ −gc
y
n0 − 1
(44)
Let f (t) denote the probability density function of the first exit time T and define hc (t; y) =
(
)
gc n0y−1 t n0y−1 . Since gc (t) is a concave function in t, it follows by Equation (9) in Jennen (1985)
that
( 2
)
hc (t; y) − th′c (t; y)
hc (t; y)
√
f (t) ≤ 2
exp −
, for any t > 0.
2t
2πt3/2
This implies that a proper N (y) can be determined by letting
( 2
)
∫
hc (t; y) − th′c (t; y)
h (t; y)
1 ∞
√
2
exp − c
dt ≤ η.
2 N (y)
2t
2πt3/2
(45)
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
Define ξ = (n0 − 1)/y, then the left-hand-side expression in (45) is simplified as
)
( 2
∫ ∞
hc (t; y)
hc (t; y) − th′c (t; y)
√
dt
exp −
2t
2πt3/2
N (y)
( 2
)
∫ ∞
gc (ξt) − (ξt)gc′ (ξt)
g (ξt)
√
=
exp − c 2
dt
2ξ t
2πt3/2 ξ
N (y)
( 2 )
∫ ∞
gc (t) − tgc′ (t)
g (t)
√
=
exp − c
dt
√
2ξt
2πt3/2 ξ
ξN (y)
( 2 ) (
) ∫ ∞
( 2 )
∫ ∞
1
gc (t)
gc (t)
gc (t)
g (t)
√
√
=−
exp −
d √
+
exp − c
dt
3/2
2ξt
2ξt
2πξ
t
ξN (y)
ξN (y) 2 2πξ t
)
(
( 2 )
∫ ∞
g (t)
g (t)
gc (ξN (y))
√
√ c 3/2 exp − c
−1+
=Φ
dt,
2ξt
2
2πξ
t
ξ N (y)
ξN (y)
39
(46)
where Φ(·) denotes the cumulative distribution function of a standard normal distribution. Notice
√
that we can always make N (y) large enough so that gc (t)/ t is monotone over [ξN (y), ∞) and
hence the changing of variables in (46) is valid.
Further, the integration term in (46) can be rewritten as
(
)
∫ ∞ √
[c + log(t + 1)](t + 1)
[c + log(t + 1)](t + 1)
√
exp −
dt
2ξt
2 2πξ t3/2
ξN (y)
)
(
∫ ∞ √
[c + log(t + 1)](t + 1)
c + log(t + 1)
√
≤
dt
exp
−
2ξ
2 2πξ t3/2
ξN (y)
(
)3/2 ∫ ∞ √
)
(
[c + log(t + 1)](t + 1)
1
c + log(t + 1)
√
≤ 1+
dt
exp −
ξN (y)
2ξ
2 2πξ (t + 1)3/2
ξN (y)
(
)3/2 ∫ ∞ √
(
) (
)
c + log(t + 1)
1
c + log(t + 1)
c + log(t + 1)
√
= 1+
exp −
d
ξN (y)
2ξ
2ξ
2πξ −1
ξN (y)
)3/2 ∫ ∞
(
ξ
1
=√
x1/2 e−x dx
1+
ξN (y)
π
[c+log(ξN (y)+1)]/(2ξ)
(
)3/2
ξ
1
=
1+
Γ ([c + log(ξN (y) + 1)]/(2ξ); 3/2) ,
2
ξN (y)
(47)
where Γ( · ; 3/2) denotes the upper incomplete gamma function with index 3/2. Based on Equations
(46) and (47), we have the truncation error
P{B(T ) ≤ −gc (T ), N (y) < T < ∞}
(
)
(
)3/2
gc (ξN (y))
ξ
1
√
≤Φ
−1+
Γ ([c + log(ξN (y) + 1)]/(2ξ); 3/2) .
1+
2
ξN (y)
ξ N (y)
(48)
Since the right-hand side in (48) decreases to 0 as the N (y) increases to the infinity, we can always
find a N (y) to meet inequality (44) for any η ∈ (0, 1) and y > 0 by solving (48) = η. It follows that
}]
[ {
(
)
n0 − 1
Y
E P B(T ) ≤ −gc
T
, N (y) < T < ∞Y
≤η
(49)
Y
n0 − 1
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
40
This implies a biased estimator
{
(
}
)
n
n0 − 1
1∑
yi
I B(Ti ) ≤ −gc
, Ti ≤ N (yi )
p̂ =
Ti
n i=1
yi
n0 − 1
(50)
{
(
)
}
with the bias less than η for the probability P B(T ) ≤ −gc n0y−1 T n0y−1 , T < ∞ , where y1 , y2 , . . .
are i.i.d. observations from Y .
Further, we implement an importance-sampling (IS) technique, which leads to a new estimator
with a smaller variance and thus a higher efficiency of estimation. Let B(t) denote a standard
Brownian motion process under P and define a new probability measure Pµ by
Lµ (t) =
dPµ
= exp(µB(t) − tµ2 /2), where µ ̸= 0.
dP
By Girsanov theorem, B̃(·) = B(t) − µt is a Brownian motion process without drift µ and unit
variance under the probability measure Pµ . Therefore, we get a new IS estimator
{
(
)
}
n
1∑
n0 − 1
yi
p̂IS =
I B̃µ(yi ) (Ti ) ≤ −gc
Ti
, Ti ≤ N (yi ) L−1
µ(yi ) (Ti )
n i=1
yi
n0 − 1
where y1 , y2 , . . . are i.i.d. observations from Y .
Next, we would like to choose the drift µ(·) so that the new estimator has a lower variance.
{
}
Define I , I B̃µ(Y ) (T ) ≤ −gc (ξ(Y )T ) /ξ(Y ), T ≤ N (Y ) L−1
µ(Y ) (T ) where ξ(y) = (n0 − 1)/y. Then
Eµ(y) [I 2 |Y = y]
[ {
}
]
(T
)
=Eµ(y) I B̃µ(y) (T ) ≤ −gc (ξ(y)T ) /ξ(y), T ≤ N (y) L−2
µ(y)
}
]
[ {
(T
)
|
Y
=Eµ(Y ) I B̃µ(Y ) (T ) ≤ −gc (ξ(Y )T ) /ξ(Y ), T ≤ N (Y ) L−2
µ(Y )
[ {
}
]
(T
)
=Eµ(y) I B̃µ(y) (T ) ≤ −gc (ξ(y)T ) /ξ(y), T ≤ N (y) L−2
µ(y)
[ {
}
(
)]
gc (ξ(y)T )
gc (ξ(y)T )
2
=Eµ(y) I B̃µ(y) (T ) ≤ −
, T ≤ N (y) exp −2µ(y)
+ T µ (y)
ξ(y)
ξ(y)
[ {
}
(
)]
gc (ξ(y)T )
gc (ξ(y)N (y))
2
≤Eµ(Y ) I B̃µ(Y ) (T ) ≤ −
, T ≤ N (y) exp −2µ(y)
+ N (y)µ (y) . (51)
ξ(y)
ξ(y)
The last inequality holds because gc (t)/t is decreasing. In the right-hand side of (51), we choose
µ∗ (y) = −gc (ξ(y)N (y))/(ξ(y)N (y)) that minimizes L−2
µ(y) (T ). It follows that
[ {
}
( 2
)]
gc (ξ(y)T )
g (ξ(y)N (y))
Eµ(y) [I |Y = y] ≤ Eµ(y) I B̃µ(y) (T ) ≤ −
, T ≤ N (y) exp − c 2
ξ(y)
ξ (y)N (y)
[ {
}]
gc (ξ(y)T )
≤ E I B̃(T ) ≤ −
, T ≤ N (y)
ξ(y)
= E[p̂2 |Y = y],
(52)
2
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
41
where the second inequality holds because of Lemma 3. Now we show that the IS estimator p̂IS has
a smaller variance the plain estimator p̂. Notice that
Var[p̂IS ] = Var[Eµ(Y ) [p̂IS |Y ]] + E[Var[p̂IS |Y ]]
= Var[Eµ(Y ) [p̂IS |Y ]] + E[Eµ(Y ) [p̂2IS |Y ]] − E[E2 [p̂IS |Y ]]
≤ Var[E[p̂|Y ]] + E[E[p̂2 |Y ]] − E[E2 [p̂|Y ]]
= Var[p̂].
This implies that the IS estimator has a smaller variance.
Using the method described, we estimate the left-hand side probability of Equation (3) for
various values of n0 and c and report them in Table 8. For any fixed α/(k − 1), users can find its
corresponding c if α/(k − 1) exists in Table 8. Otherwise, users can start with finding two constants
in the table such that α/(k − 1) lies between them and then approximate c using the interpolation
method.
References
Basu, D. 1955 On statistics independent of a complete sufficient statistic. Sankhyā: The Indian Journal of
Statistics 377–380.
Batur, D., S. H. Kim. 2006. Fully sequential selection procedures with a parabolic boundary. IIE Transactions
38(9) 749–764.
Bechhofer, R. E. 1954. A single-sample multiple decision procedure for ranking means of normal populations
with known variances. The Annals of Mathematical Statistics 16–39.
Bechhofer, R. E., T. J. Santner, D. M. Goldsman. 1995. Design and analysis of experiments for statistical
selection, screening, and multiple comparisons. New York: John Wiley & Sons.
Branke, J., S. E. Chick, C. Schmidt. 2007. Selecting a selection procedure. Management Science 53(12)
1916–1932.
Bubeck, S., N. Cesa-Bianchi. 2012. Regret analysis of stochastic and nonstochastic multi-armed bandit
problems. Foundations and Trends in Machine Learning 5(1) 1–122.
Chick, S. E., K. Inoue. 2001. New two-stage and sequential procedures for selecting the best simulated
system. Operations Research 49(5) 732–743.
Chick, S. E., P. I. Frazier. 2012. Sequential sampling with economics of selection procedures. Management
Science 58(3) 550–569.
Clark, G. M., W. Yang. 1986. A bonferroni selection procedure when using commom random numbers with
unknown variances. Proceedings of the 1986 Winter Simulation Conference (WSC) 313–315.
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
42
Dudewicz, E. J. 1969. An approximation to the sample size in selection problems. Annals of Mathematical
Statistics 40(2) 492–497.
Durbin, J. 1985. The first-passage density of a continuous gaussian process to a general boundary. Journal
of Applied Probability 99–122.
Durrett, R. 2010. Probability: Theory and Examples, vol. 3. Cambridge university press.
Egoroff, D. T. 1911. Sur les suites de fonctions mesurables. CR Acad. Sci. Paris 152 244–246.
Fan, W., L. J. Hong. 2014. A frequentist selection-of-the-best procedure without indifference zone. Proceedings of the 2014 Winter Simulation Conference (WSC) 3737–3748.
Frazier, P. I. 2014. A fully sequential elimination procedure for indifference-zone ranking and selection with
tight bounds on probability of correct selection. Operations Research 62(4) 926–942.
Gupta, S. S. 1956. On a decision rule for a problem in ranking means. Ph.D. thesis, University of North
Carolina. Chapel Hill.
Gupta, S. S. 1965. On some multiple decision (selection and ranking) rules. Technometrics 7(2) 225–245.
Hong, L. J. 2006. Fully sequential indifference-zone selection procedures with variance-dependent sampling.
Naval Research Logistics 53(5) 464–476.
Hong, L. J., B. L. Nelson. 2005. The tradeoff between sampling and switching: New sequential procedures
for indifference-zone selection. IIE Transactions 37(7) 623–634.
Hong, L. J., B. L. Nelson. 2007. Selecting the best system when systems are revealed sequentially. IIE
Transactions 39(7) 723–734.
Jamieson, K., R. Nowak. 2014. Best-arm identification algorithms for multi-armed bandits in the fixed
confidence setting. The 48th Annual Conference on Information Sciences and Systems (CISS) IEEE,
1–6.
Jennen, C., H. R. Lerche. 1981. First exit densities of brownian motion through one-sided moving boundaries.
Zeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete 55(2) 133–148.
Jennen, C. 1985. Second-order approximations to the density, mean and variance of Brownian first-exit
times. The Annals of Probability 13(1) 126–144.
Jennison, C., I. M. Johnston, B. W. Turnbull. 1980. Asymptotically optimal procedures for sequential
adaptive selection of the best of several normal means. Technical Report, Dept. of ORIE, Cornell Univ.,
Ithaca, New York .
Karatzas, I. A., S. E. Shreve. 1991. Brownian Motion and Stochastic Calculus, vol. 113. Springer, New York.
Karnin, Z., K. Tomer, S. Oren, 2013. Almost optimal exploration in multi-armed bandits. Proceedings of
the 30th International Conference on Machine Learning (ICML-13) 1238–1246.
Kim, S. H., B. L. Nelson. 2001. A fully sequential procedure for indifference-zone selection in simulation.
ACM Transactions on Modeling and Computer Simulation (TOMACS) 11(3) 251–273.
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
43
Kim, S. H., B. L. Nelson. 2006a. Selecting the best system. Handbooks in Operations Research and Management Science: Simulation 13 501–534. Elsevier.
Kim, S. H., B. L. Nelson. 2006b. On the asymptotic validity of fully sequential selection procedures for
steady-state simulation. Operations Research 54(3) 475–488.
Ni, E. C., S. R. Hunter, S. G. Henderson, H. Topaloglu. 2012. Exploring bounds on ambulance deployment
policy performance. Proceedings of the 2012 Winter Simulation Conference (WSC) 1–12.
Ni, E. C., S. G. Henderson, S. R. Hunter. 2014. A comparison of two parallel ranking and selection procedures.
Proceedings of the 2014 Winter Simulation Conference (WSC) 3761–3772.
Panchapakesan, S., S. S. Gupta, J. Yackel. 1971. On a subset selection procedure for the most probable
event in a multinomial distribution. Statistical Decision Theory and Related Topics 275–298.
Paulson, E. 1964. A sequential procedure for selecting the population with the largest mean from k normal
populations. The Annals of Mathematical Statistics 35(1) 174–180.
Perng, S. K. 1969. A comparison of the asymptotic expected sample sizes of two sequential procedures for
ranking problem. The Annals of Mathematical Statistics 40(6) 2198–2202.
Pichitlamken, J., B. L. Nelson, L. J. Hong. 2006. A sequential procedure for neighborhood selection-of-thebest in optimization via simulation. European Journal of Operational Research 173(1) 283–298.
Rinott, Y. 1978. On two-stage selection procedures and related probability-inequalities. Communications in
Statistics-Theory and methods 7(8) 799–811.
Siegmund, D. 1967. Some one-sided stopping rules. The Annals of Mathematical Statistics 38(6) 1641–1646.
Sullivan, D. W., J. R. Wilson. 1989. Restricted subset selection procedures for simulation. Operations
Research 37(1) 52–71.
Wang, H., S. H. Kim. 2013. Reducing the conservativeness of fully sequential indifference-zone procedures.
IEEE Transactions on Automatic Control 58(6) 1613–1619.
Whitt, W. 2002a. Stochastic-process limits: an introduction to stochastic-process limits and their application
to queues. Springer, New York.
Whitt, W. 2002b. Internet supplement to stochastic-process limits (2002). URL http://www. columbia. edu/˜
ww2040/supplement.html .
Fan, Hong and Nelson: Indifference-Zone-Free Selection of the Best
Article submitted to Operations Research; manuscript no. (Please, provide the manuscript number!)
44
Table 8
Table of β for different values of c and n0 when gc (t) =
c
4.0
4.2
4.4
4.6
4.8
5.0
5.2
5.4
5.6
5.8
6.0
6.2
6.4
6.6
6.8
7.0
7.2
7.4
7.6
7.8
8.0
8.2
8.4
8.6
8.8
9.0
9.2
9.4
9.6
9.8
10.0
10.2
10.4
10.6
10.8
11.0
11.2
11.4
11.6
11.8
12.0
12.2
12.4
12.6
12.8
13.0
13.2
13.4
13.6
13.8
14.0
14.2
14.4
14.6
14.8
15.0
10
0.1042
0.0990
0.0945
0.0905
0.0862
0.0822
0.0786
0.0750
0.0718
0.0691
0.0657
0.0632
0.0607
0.0581
0.0556
0.0538
0.0519
0.0495
0.0479
0.0458
0.0443
0.0424
0.0410
0.0393
0.0382
0.0369
0.0357
0.0341
0.0330
0.0319
0.0309
0.0295
0.0289
0.0280
0.0269
0.0259
0.0252
0.0243
0.0236
0.0228
0.0221
0.0215
0.0207
0.0199
0.0197
0.0190
0.0184
0.0178
0.0173
0.0168
0.0163
0.0158
0.0153
0.0150
0.0145
0.0141
15
0.0798
0.0801
0.0750
0.0711
0.0668
0.0631
0.0596
0.0560
0.0534
0.0508
0.0485
0.0456
0.0431
0.0410
0.0392
0.0368
0.0352
0.0335
0.0315
0.0301
0.0291
0.0271
0.0261
0.0249
0.0235
0.0224
0.0217
0.0204
0.0195
0.0186
0.0180
0.0172
0.0164
0.0156
0.0150
0.0141
0.0137
0.0129
0.0125
0.0121
0.0114
0.0110
0.0105
0.0101
0.0098
0.0094
0.0089
0.0086
0.0083
0.0079
0.0077
0.0073
0.0070
0.0067
0.0065
0.0061
20
0.0744
0.0697
0.0652
0.0613
0.0572
0.0538
0.0506
0.0476
0.0442
0.0417
0.0392
0.0367
0.0348
0.0327
0.0305
0.0287
0.0271
0.0257
0.0249
0.0227
0.0215
0.0202
0.0191
0.0182
0.0174
0.0164
0.0152
0.0145
0.0137
0.0131
0.0124
0.0117
0.0113
0.0107
0.0099
0.0096
0.0090
0.0042
0.0082
0.0077
0.0073
0.0070
0.0066
0.0064
0.0059
0.0056
0.0053
0.0051
0.0048
0.0047
0.0045
0.0042
0.0041
0.0038
0.0037
0.0035
n0
25
0.0686
0.0639
0.0594
0.0552
0.0515
0.0479
0.0450
0.0419
0.0389
0.0367
0.0340
0.0320
0.0298
0.0279
0.0261
0.0246
0.0230
0.0214
0.0200
0.0188
0.0179
0.0167
0.0155
0.0145
0.0140
0.0130
0.0121
0.0115
0.0107
0.0101
0.0095
0.0091
0.0085
0.0081
0.0074
0.0071
0.0066
0.0064
0.0060
0.0056
0.0054
0.0050
0.0047
0.0043
0.0042
0.0040
0.0038
0.0036
0.0033
0.0032
0.0031
0.0029
0.0027
0.0026
0.0025
0.0023
30
0.0647
0.0594
0.0552
0.0512
0.0478
0.0442
0.0411
0.0380
0.0359
0.0330
0.0306
0.0287
0.0266
0.0248
0.0231
0.0216
0.0201
0.0189
0.0175
0.0163
0.0152
0.0142
0.0134
0.0124
0.0115
0.0110
0.0102
0.0094
0.0089
0.0085
0.0078
0.0074
0.0069
0.0066
0.0059
0.0058
0.0054
0.0049
0.0047
0.0044
0.0041
0.0039
0.0037
0.0033
0.0032
0.0030
0.0029
0.0027
0.0026
0.0024
0.0023
0.0021
0.0020
0.0020
0.0018
0.0017
√
[c + log(t + 1)](t + 1)
35
0.0613
0.0569
0.0524
0.0482
0.0449
0.0412
0.0383
0.0354
0.0328
0.0305
0.0282
0.0267
0.0245
0.0227
0.0211
0.0198
0.0182
0.0169
0.0158
0.0145
0.0135
0.0127
0.0118
0.0111
0.0101
0.0094
0.0088
0.0083
0.0077
0.0072
0.0067
0.0063
0.0058
0.0054
0.0050
0.0047
0.0044
0.0041
0.0039
0.0036
0.0035
0.0032
0.0030
0.0027
0.0026
0.0025
0.0023
0.0021
0.0020
0.0018
0.0019
0.0017
0.0016
0.0014
0.0014
0.0013
40
0.0593
0.0542
0.0503
0.0463
0.0428
0.0393
0.0366
0.0338
0.0313
0.0289
0.0264
0.0248
0.0226
0.0212
0.0196
0.0181
0.0166
0.0157
0.0145
0.0134
0.0124
0.0116
0.0107
0.0099
0.0093
0.0085
0.0079
0.0074
0.0069
0.0064
0.0058
0.0054
0.0052
0.0047
0.0045
0.0042
0.0038
0.0036
0.0033
0.0031
0.0029
0.0027
0.0025
0.0024
0.0023
0.0021
0.0019
0.0018
0.0016
0.0016
0.0015
0.0014
0.0012
0.0012
0.0011
0.0010
Note: All the estimated values reported in this table are within 0.0001 of the true values with 95% confidence level.