1 Introduction

Let \(F_{\theta }(x)\) be a family of distributions indexed by its mean \(\theta .\) Suppose there are n populations, and that each new observation from population i is independent of all previous observations and is the value of a random variable with distribution \(F_{\theta _i },\) where \(\theta _1, \ldots , \theta _n\) are unknown. Furthermore, suppose that our objective is to decide which population has the largest mean. It is supposed that a decision is made at each stage as to which population to next take an observation from, with the decision made according to some rule which eventually calls for stopping and declaring which population has the largest mean. Our objective is to determine a rule that makes a decision in a relatively small expected time, subject to the condition that its probability of making a correct choice is at least \(\alpha .\) We will study such models both when the population distributions are Bernoulli and when they are normal with a fixed variance. In the Bernoulli case, we suppose that the result of an observation is either a success or a failure, and the objective is to find the population having the largest success probability.

These type of models have many applications. Foremost is probably in clinical trials to determine which of several medical approaches (e.g., drugs, treatments, and procedures) yields the best results. Here, a population would refer to a particular approach, with its use resulting in either a success (suitably defined) or not. Another application is to online advertising, where a decision maker is trying to decide which of n different advertisements to utilize. For instance, the advertisements might relate to army recruitment, and a success might refer to a subsequent clicking on the advertisement to obtain additional information. Another application is to choose among different methods for teaching a particular skill. Each day, a method can be used on a group of students, with the students being tested at the end of the day with each test resulting in a score which would be pass (1) or fail (0) in the Bernoulli case, and numerical in the normal case.

It should be noted that these problems have been studied for quite some time. However, the early work, such as in [3, 4, 8, 9, 11, 15, 16], was done under the assumption that the difference between the largest and second largest population mean was at least some known positive value. More recent work, such as [1, 5,6,7, 10, 13], does not make this assumption. What primarily distinguishes our models from others considered in the literature is that we take a Bayesian approach that supposes the unknown means are the respective values of independent and identically distributed (iid) random variables having a specified prior distribution F. Moreover, we present numerical evidence that the rules obtained when we assume that F is the uniform (0, 1) distribution tend to perform well for most sets of success probabilities.

In Sect. 2, we consider the Bernoulli case where we want to find rules whose implementation results in a relatively small expected number of observations needed before a decision is made, subject to the condition that the rule results in the probability of a correct choice being at least some specified value \(\alpha \). We reconsider the classical “vector at a time" (VT) rule, which is such that each alive population is sampled once in each round. At the end of a round, any population whose cumulative number of successes is k less than another is no longer alive. The process ends when only one population is alive. The appropriate value of k that results in the probability of a correct choice being at least \(\alpha \) was determined in [15] and [16] under the assumption that the difference between the largest and second largest population mean is at least some specified positive value \(d > 0.\) We show, in this section, how this rule can be implemented and analyzed in our Bayesian setting. In particular, we present a lower bound as well as an accurate approximation for the probability that VT using critical value k makes the correct decision, as well as an approximation for the mean number of observations it takes. These bounds and approximations can be evaluated by a simple simulation, with each simulation run requiring either 2 or 3 random numbers. We also present some numerical evidence, indicating that the version of this rule that results when we assume a uniform (0, 1) prior appears to outperform more recently proposed rules, even when the means do not come from the uniform prior.

In Sect. 3, we consider improving the VT rule by allowing for early elimination if a population is j behind another after j rounds. (That is, if one population has had 0 successes and another j successes in their first j observations, then the former is no longer alive.) We show how to determine the probability that the population with largest mean is eliminated early, as well as the mean number of the non-best populations that are eliminated early.

In Sect. 4, we consider another classical rule: “play the winner" (PW). Our variant of the PW rule is such that in all but the final round each alive population is continually sampled until it has a failure, where a population is no longer alive if after a round its cumulative number of successes is at least k smaller than that of some other arm. This variant differs from the classical model of [15] which declared a population dead if at some point—not necessarily at the end of a round—it has k fewer successes than another population. We show how to analyze this rule in the Bayesian setting.

In Sect. 5, we consider the case of normal populations, where population distributions are all normal distributions with some fixed variance \(\sigma ^2,\) and where the standard normal is the prior distribution on the means. Among other things, by using that a normal conditioned to be positive has an increasing failure rate, we improve upon known bounds for the probability that a random walk of normal random variables reaches a before falling as low as \(-b\) for given positive numbers ab.

Section 6 gives the paper’s conclusions.

2 The Vector at a Time Rule in the Bernoulli Case

Suppose there are n Bernoulli populations with respective means \(p_1, \ldots , p_n\), and that at each stage we are allowed to make an observation from a population of our choice, stopping these observations at some point and making a decision as to which population has the largest mean. As noted earlier, we suppose that \(p_1, \ldots , p_n\) are the respective values of iid random variables having a specified distribution F. Subject to the constraint that the policy used results in the probability of a correct choice being at least \(\alpha ,\) the objective is to choose a policy whose mean number of observations is relatively small.

Definition

The vector at a time (VT) rule, introduced in [3], is defined as follows. For a given positive integer k, depending on the desired accuracy \(\alpha ,\) the policy is as follows.

  • Initially, all populations are alive

  • A round consists of a single observation from each alive population.

  • At the end of a round, a population is no longer alive if its cumulative number of successes is k less than that of another population.

  • If only one population is alive, stop and declare it best. Otherwise, perform a new round.

Let \(p_{[1]}> p_{[2]}> \ldots > p_{[n]}\) be the ordered values of the unknown means \(p_1, \ldots , p_n,\) and let C be the event that the correct choice is made. Under the assumption that there is a known positive value d such that \( p_{[1]} - p_{[2]} > d,\) it was shown in [15] how to determine k so that \(P(C) \ge \alpha .\) We now show how this can be done when \(p_1, \ldots , p_n\) are the values of independent and identically distributed random variables having distribution F.

Notation: We use the notation \(I\{A\}\) to be the indicator of the event A,  equal to 1 if A occurs and to 0 otherwise. Also, we use the notation \(X =_{st} Y\) to indicate that X and Y have the same distribution.

Lemma 1

Let \( X_1, \ldots , X_{n}\) be independent and identically distributed random variables having distribution F, and let \(U, U_1, \ldots , U_{n}\) be independent uniform (0, 1) random variables. Then,

$$\begin{aligned} \quad \max _i X_i =_{st} F^{-1}(U^{1/n}) \end{aligned}$$

Proof

 

$$\begin{aligned} \max _i X_i&=_{st}&\max (F^{-1}(U_1), \ldots , F^{-1}(U_n) ) \\ {}= & {} {} F^{-1}(\max _{i=1, \ldots , n} U_i ) \\ {}&=_{st}&F^{-1} (U^{1/n} ) \end{aligned}$$

We now show how to bound P(C), the probability that the correct population is chosen. To begin, suppose that in each round we take an observation from each population, even those that may be dead. Let 0 be the best population, namely the one with largest mean, and randomly number the others as population \(1, \ldots , n-1.\) Imagine that the best population is playing a “gambler’s ruin game" with each of the others, with the best one beating population i if the difference of the cumulative number of wins of the best to that of i hits the value k before \(-k.\) Let \(B_i\) be the event that the best population beats \(i, i = 1, \ldots , n-1,\) and note that the best population will be chosen if it wins all of its games. That is, if we let \(\;B \equiv B_1B_2 \cdots B_{n-1} ,\;\) then \(\;B \subset C,\) giving that

$$\begin{aligned} P(C) \ge P(B) \end{aligned}$$

\(\square \)

Lemma 2

\(P(B) \ge (P(B_1))^{n-1}\)

Proof

Let \(U_0, U_1, \ldots , U_{n-1}\) and \(U_{i,j}, i = 0, 1, \ldots , n-1, j \ge 1\) all be independent uniform (0, 1) random variables. Let \(X_0 = F^{-1}(U_0^{1/n}),\; X_i = F^{-1}((1-U_i) U_0^{1/n}), i = 1, \ldots , n-1.\) Using Lemma 1 along with the fact that conditional on the maximum, call it M,  of n independent uniform (0, 1) random variables, the \(n-1\) of these uniforms whose values are less than M are distributed as independent uniform (0, M) random variables, it follows that the joint distribution of \(X_0, X_1, \ldots , X_{n-1}\) is exactly that of the mean of the best population, followed by the means of the other \(n-1\) populations in a random order.

Let \(I_{0,j} = I\{1-U_{0,j} < X_0\}, j \ge 1,\) and \(I_{i,j} = I\{U_{i,j} < X_i\}, i = 1, \ldots , n-1, j \ge 1.\) Note that \(I_{i,j}\) has the distribution of the \(j^{th}\) observation of population \(i,\; i = 0, \ldots , n-1, j \ge 1,\) and also that \(I_{0,j}\) is increasing in \(U_{0,j}\) whereas \(I_{i,j}\) is decreasing in \(U_{i,j}, i \ge 1.\) Because \(X_i\) is decreasing in \(U_i\) for \(i>0\), it consequently follows that, conditional on \(U_0,\) the indicator variables \(I\{B_1 \}, \ldots , I\{B_{n-1}\}\) are all increasing functions of the independent random variables \(U_1, \ldots , U_{n-1}, U_{i,j} , i = 0, \ldots , n-1, j \ge 1.\) Consequently, given \(U_0,\) the indicators \(I\{B_1 \}, \ldots I\{B_{n-1}\}\) are associated, implying that

$$\begin{aligned} P(B |U_0) \ge \prod _{i=1}^{n-1} P(B_i|U_0) \end{aligned}$$

which, by symmetry gives

$$\begin{aligned} P(B|U_0) \ge (P(B_1|U_0))^{n-1}. \end{aligned}$$

Taking expectations gives

$$\begin{aligned} P(B)\ge & {} E[ (P(B_1|P_0))^{n-1}] \\\ge & {} (E[ P(B_1|P_0)])^{n-1} \\= & {} (P(B_1))^{n-1} \end{aligned}$$

where the last inequality follows from Jensen’s inequality. \(\square \)

To obtain an upper bound on P(B),  let \(B^*\) be the event that the population with largest mean wins its gambler’s ruin game against the population with the second largest mean. Because \(B \subset B^*\), we have

Lemma 3

\(P(B) \le P(B^*) \)

Remark

It is possible for the best arm to be chosen even if it does not win all its games. Indeed, this will happen if the best arm loses to an arm that at an earlier time had become dead. However, it is intuitive that this event has a very small probability of occurrence. Consequently, \(P(C) \approx P(B).\)

To compute \(P(B_1)\) and \(P(B^*)\), we will use simulation with a conditional expectation estimator. To begin, note that if populations with known probabilities x and y play a game that ends when one has k more wins than the other, then the probability that the one with probability x wins is the probability that a gambler, starting with fortune k, who wins each game with probability \( p = \frac{ x(1-y)}{x(1-y) + y(1-x)}\) will reach a fortune of 2k before 0. But, with

$$\begin{aligned} r \equiv \frac{1-p}{p} = \frac{ y(1-x)}{x(1-y) } \end{aligned}$$

this gambler’s ruin probability is

$$\begin{aligned} P(x\;\text{ wins}) = \frac{ 1- r^k }{ 1- r^{2k} } = \frac{ 1}{ 1 + r^k } \end{aligned}$$
(1)

Also, using known results from the gambler’s ruin problem along with Wald’s equation (to account for the fact that not every round leads to a gain or a loss) it follows that the mean number of plays is

$$\begin{aligned} E[\text{ number } \text{ of } \text{ plays}] = \frac{ k( 1-r^k )}{ (r^k + 1)(x-y)} \end{aligned}$$
(2)

Proposition 1

Let U and V be independent uniform (0, 1) random variables, and let

$$\begin{aligned} X= & {} F^{-1}( U^{1/n} )\,, \;\; Y = F^{-1}( U^{1/n} V) \,, \;\; W = F^{-1}( U^{1/n} V^{1/(n-1)}) \\ R= & {} \frac{ Y( 1 - X) }{ X(1-Y) } \,, \quad S = \frac{ W( 1 - X) }{ X(1-W) } \end{aligned}$$

then

$$\begin{aligned} P(B_1) = E\left[ \frac{ 1}{ 1 + R^k } \right] \,, \quad P(B^*) = E\left[ \frac{ 1}{ 1 + S^k } \right] \end{aligned}$$

Proof

The result follows from Eq. (1) upon using that the joint distribution of XY is that of the largest and a random one of the other means, and the joint distribution of XW is that of the largest and second largest means. \(\square \)

Letting N be the number of observations, we can approximate E[N] by approximating the mean number of plays of each of the non-best populations by the mean number of plays in their game against the best population, and approximating the mean number of plays of the best population by the mean number of plays in its game against the second best one. Hence, using Eq. (2) we see that

$$\begin{aligned} E[N] \approx A \equiv (n-1) E\left[ \frac{ k( 1-R^k )}{ (R^k + 1)(X-Y)} \right] + E\left[ \frac{ k( 1-S^k )}{ (S^k + 1)(X- W)} \right] \end{aligned}$$
(3)

Using Proposition 1 and Eq. (3) enables us to efficiently estimate \(P(B_1), P(B^*),\) and A by a simulation. Indeed, a simulation of 1, 000, 000 runs, yielded that when \(n=10\) and \(F(x) = x, 0 \le x \le 1,\) the value \(k=50\) resulted in the estimates \(P(B_1)^{n-1} = .9885,\; P(B^*) = .9889 ,\) and that \(A \approx 5460.4\), with an estimate of the standard deviation of the estimator of A being 26.7. Thus, \(.9885< P(B) < .9889 ,\) and \(E[N] \approx 5460.539.\) A much more time consuming simulation (the preceding takes a fraction of a second) consisting of 1,000,000 runs, each run generating a random variable distributed as N, yielded that \(P(C) \approx 0.9886, \; E[N] \approx 5466.318,\, \) with the standard deviation of the estimate of E[N] being 17.34.

For another illustration of the utility of the approximations of P(C) and of the mean number of observations, suppose \(n=5\) and VT with \(k=10\) is used. A simulation based on Proposition 1 and Eq. (3) yielded that

$$\begin{aligned} 0.9523404< P(B) < 0.9561526, \quad A \approx 358.398320 \end{aligned}$$

with a standard deviation of the estimate of A being 0.8256477. A simulation with 500, 000 runs, with each run generating the value of N, yielded that

$$\begin{aligned} P(C) = 0.9540, E[N] = 358.3993, \text{ sd } = 1.156 \end{aligned}$$

where sd refers to the standard deviation of the estimator of E[N].

The following example compares the performance of VT with some recently proposed rules (Table 1).

Example 1

Comparison with Recent Literature One might hope that the vector at a time rule assuming a uniform (0, 1) prior assumption performs well for any set of probabilities \(p_1, \ldots , p_n.\) The following compares its performance with some recent algorithms—with names such as track and stop, Chernoff racing, and Kullback–Leibler racing (see [5]. [10], and [6]. These algorithms, some of which have some asymptotic optimality features as the desired accuracy goes to 1, solve optimization problems to determine which population to next observe (and, consequently, are much more difficult to implement than is the VT rule). Table 1 is taken from [7]. It gives the results of 5 of these algorithms for two cases: the first case having \(n=4\) with probabilities: (0.5, 0.45, 0.43, 0.4),  and the second having \(n=5\) with probabilities (0.3, 0.21, 0.20, 0.19, 0.18). The parameters of the algorithms are chosen to guarantee at least 90 percent accuracy. (TS1 and TS2 refer to two variants of the track and stop algorithm; CR refers to the Chernoff racing algorithm, and KL1 and KL2 refer to two variants of the Kullback–Leibler racing algorithm.

Table 1 Algorithm comparison

Because our algorithm assumes knowledge of a prior distribution, in cases where there is no reason to assume that we know what the prior is, it seems reasonable to assume a uniform (0, 1) prior and choose a larger accuracy than is actually desired. So suppose we do so and require an accuracy, under a uniform (0, 1) prior, of 99 percent. When \(n=4\), the vector at a time algorithm assuming a uniform prior requires \(k= 42,\) and when \(n=5\), if requires \(k=47.\) Simulation, using the probabilities in each case, yields that the average number of observations needed in case 1 is 2738 with the correct decision being made with probability 0.9999;  whereas the average number of observations needed in case 2 is 2372 with the correct decision being made with probability 0.9999. If we were to be less conservative and only require 97 percent accuracy when assuming the uniform prior, then the required value of the vector at a time rule is \(k=15\) when \(n=4\) and \(k=16\) when \(n=5.\) Simulation yields that the average number of observations needed in case 1 is 905 with the correct decision being made with probability 0.9999;  whereas the average number of observations needed in case 2 is 832 with the correct decision being made with probability 0.998. Thus, the vector at a time algorithm significantly outperforms the newer algorithms. (Though to be fair we should mention that, under the uniform (0, 1) prior, \(k=5\) is sufficient when either \(n=4\) or \(n=5\) to obtain 90 percent accuracy. Using \(k=5\) in Case 1 yields that the average number of observations needed is only 166, but the accuracy is .601. Using \(k=5\) in Case 2 yields that the average number of observations needed is 213.2, with accuracy .81.)

Remark

The preceding example is very interesting in that it indicates that the vector at a time algorithm that assumes a uniform (0, 1) prior can significantly outperform the newer algorithms even in cases where the probabilities are highly unlikely to have come from this prior. Thus, while we are not claiming that there are not cases where assuming a uniform prior will lead to a poor result (for instance, for any value of k chosen, if all the \(p_i\) are approximately equal, then the VT procedure will yield \(P(C) \approx 1/n,\)) we do feel that it will typically perform quite well.

Remark on Variance Reduction:

In practice, we observe that the number of plays using VT may have a large variance. In the case where F is the uniform (0, 1) distribution, we can reduce the variance of a simulation estimator that on each run generates the value of N by using \(Y = \frac{1}{P_1 (1 - P_2)}\) as a control variable, where \(P_1\) and \(P_2\) are the means of the best and second best arms. That is, if let T denote the raw estimator, then the new estimator is

$$\begin{aligned} T + c(Y - E[Y]) \end{aligned}$$

where the variance is minimized when \(c = -\frac{\text{ Cov }(T, Y)}{\text{ Var }(Y)}\). To obtain the mean value of the control variable, we condition on \(P_2\),

$$\begin{aligned} E\left[ \frac{1}{P_1(1 - P_2)}\right]= & {} E\left[ E\left[ \frac{1}{P_1(1 - P_2)} | P_2\right] \right] \\= & {} E\left[ \frac{1}{1-P_2} E\left[ \frac{1}{P_1}|P_2\right] \right] \\= & {} E\left[ \frac{-\log (P_2)}{(1 - P_2)^2}\right] \qquad \text{ because }\;\;P_1|P_2 \sim \text{ unif }(P_2, 1) \\= & {} n(n-1) \int _0^1 \frac{-x^{n-2}\log (x)}{(1-x)} dx \\\approx & {} n(n-1) \frac{1}{r} \sum _{i = 1} ^ r h\left( \frac{i - 0.5}{r}\right) \end{aligned}$$

where r is a large integer, and \(\, h(x) = \frac{-x^{n-2}\log (x)}{(1-x)} .\,\) The values of \(\text{ Cov }(T, Y)\) and \(\text{ Var }(Y)\) can be estimated from the simulation, and these can then be used to determine c. In our numerical examples, we observe that the variance is reduced by between 50 and 60 percent.

3 VT with Early Elimination

Suppose we use VT but with an early elimination possibility in that if an arm is j behind after the first j rounds (that is, if the arm had all failures in the first j rounds while another arm had all successes), then that arm is eliminated. To see by how much that can reduce the accuracy of VT, let us compute P(L),  where L is the event that the best arm is eliminated early. Let 0 be the best population, let \(1, \ldots , n-1\) be the other populations in random order, and let \(X_0, \ldots , X_{n-1} \) be their respective success probabilities. With \(U, U_1, \ldots , U_{n-1}\) being iid uniform (0, 1) random variables, note that \((X_0, \ldots , X_{n-1}) =_{st} ( F^{-1}( W), F^{-1}( WU_1), \ldots , F^{-1}( W U_{n-1} ) ) \) where \(W = U^{1/n}.\) Consequently, with \((X_0, \ldots , X_{n-1}) = ( F^{-1}( W), F^{-1}( WU_1), \ldots , F^{-1}( W U_{n-1} ) ) ,\) we have

$$\begin{aligned} P(L) = E\left[ (1-X_0)^j (1 - \prod _{i \ne 0} (1 - X_i^j) \right] \end{aligned}$$

Let us now consider the expected number of non-best populations that are eliminated early. Letting \(I_k\) be the indicator or the event that population k is eliminated early, we have for \(k \ge 1\)

$$\begin{aligned} E[I_k] = E\left[ (1-X_k)^j(1 - (1-X_0^j) \prod _{i \ne 0, k} (1 - X_i^j)\right] \end{aligned}$$

Hence, with \(N^*\) being the number of non-best populations that are eliminated early, we have

$$\begin{aligned} E[N^*] = (n-1) E\left[ (1-X_1)^j(1 - (1-X_0^j) \prod _{i =2}^{n} (1 - X_i^j)\right] \end{aligned}$$

P(L) and \(E[N^*]\) are easily evaluated by a simulation.

The formulas for P(L) and \(E[N^*]\) considerably simplify when F is the uniform (0, 1) distribution. Let \(N_i, i = 0, \ldots , n-1,\) be the number of successes of population i in the first j rounds. Because \(F^{-1}(x) = x, \) we obtain when \(i \ge 1\) that

$$\begin{aligned} P(N_i = j|W) = E[ (U_i W)^j| W] = \frac{ W^j}{j+1} \end{aligned}$$

Using that \(N_0, \ldots , N_{n-1}\) are conditionally independent given W, the preceding gives

$$\begin{aligned} P(L|W) = (1 - W)^j \left( 1 - \left( 1 - \frac{ W^j}{j+1} \right) ^{n-1} \right) \end{aligned}$$

Taking expectations gives

$$\begin{aligned} P(L) = E\left[ (1 - W)^j \left( 1 - \left( 1 - \frac{ W^j}{j+1} \right) ^{n-1} \right) \right] \end{aligned}$$

where \(W = U^{1,n}.\) Let us now consider \(E[N^*]\). Again using that \(N_0, \ldots , N_{n-1}\) are conditionally independent given W, we have

$$\begin{aligned} P(N_{1} = 0, \; \max _{i\ne 1 }N_i = j |W)= & {} P(N_{1} = 0| W ) \left( 1 - \prod _{i \ne 1} P(N_i \ne j |W) \right) \end{aligned}$$
(4)

Now,

$$\begin{aligned} P(N_{1} = 0| W )= & {} E[ (1 - U_{1} W)^j|W] \\= & {} \int _0^1 (1 - xW)^j \, dx \\= & {} \frac{1 - (1-W)^{j+1} }{ (j+1) W} \end{aligned}$$

Also,

$$\begin{aligned} \prod _{i \ne 1}P(N_i \ne j |W)= & {} (1 - W^j) (E[ 1- (U_1W)^j |W])^{n-2} \\= & {} (1 - W^j) \left( 1 - \frac{W^j}{j+1} \right) ^{n-2} \end{aligned}$$

Hence, Eq. (4) yields

$$\begin{aligned} E[N^*] = (n-1) E \left[ \frac{1 - (1-W)^{j+1} }{ (j+1) W} \, \left( 1 - (1 - W^j) \left( 1 - \frac{W^j}{j+1} \right) ^{n-2} \right) \right] \end{aligned}$$

where \(W = U^{1/n}.\)

Thus, when \(F(x) = x,\) both P(L) and \(E[N^]\) are one-dimensional integrals, easily evaluated by numerical methods.

Example 2

The following are the values of P(L),  the probability that the best population is eliminated early, and \(E[N^*],\) the mean number of non-best populations that are eliminated early, for a variety of values of n and j when F is the Uniform (0, 1) distribution.

n

j

P(L)

\(E[N^*]\)

5

2

.02053

1.317

3

.00336

0.851

4

.00059

0.590

5

.00011

0.432

10

2

.01278

3.234

3

.00201

2.310

4

.00033

1.731

5

.00006

1.343

20

2

.00429

6.659

3

.00053

4.978

4

.00007

3.942

5

.00001

3.229

For the cases considered in the preceding table, for a fixed j the probability that the best population is eliminated early decreases in the number of populations n. Intuitively, the reason for this is that although it becomes much more likely that at least one of the non-best populations has j successes in its first j trials as n increases, because the success probability of the best population is distributed as the maximum of n independent uniform (0, 1) random variables—and thus stochastically increases in \(n---\) the larger n is, the less likely it is that the first j observations of the best population will all be failures. \(\square \)

Let \(P_k(C)\) be the probability of a correct choice when using VT with critical value k,  and suppose \(P_{k-1}(C)< \alpha < P_{k}(C).\) The randomized rule that chooses VT with critical value k with probability \(p=\frac{\alpha - P_{k-1}(C) }{ P_{k}(C) - P_{k-1}(C) } \) or VT with critical value \(k-1\) with probability \(1-p\) will yield the correct choice with probability \(\alpha .\) Another possibility is to use VT along with early elimination parameter \(j^*,\) where \(j^*\)is the smallest value j for which using VT with critical value k along with early elimination if behind by j after the first j rounds results in a correct choice with probability at least \(\alpha .\) (Of course, we could use a policy that randomizes between VT with critical value k and early elimination at \(j^* - 1\) and VT with critical value k and early elimination at \(j^* .\)) The following is an example where randomizing among VT rules results in a smaller mean number of observations than does VT with early elimination.

Example 3

Suppose \(n= 5,\,\alpha = .95\) and \(F(x) = x, \,0 \le x \le 1.\) The following simulated results were based on 500, 000 runs. (The term sd refers to the standard deviation of the estimator of E[N].)

figure a

Consequently, randomizing among VT with \(k=9\) and \(k=10\) to obtain \(P(C) = .95\) results in the mean number of observations being \((2/3)313.64 + (1/3)358.40= 328.56,\) which is smaller than what can be obtained with VT with early elimination. (It is also better than the newer proposed rules. Of these, the Chernoff bound algorithm performs best, giving a mean number of 423.4 with accuracy 0.953.)

4 Play the Winner Rule

Another older rule that can also be utilized in the Bayesian setting is the play the winner (PW) rule, which in each but the last round continues to sample from each alive population until it has a failure. Our variation of PW, which is somewhat different than what has been previously considered, is as follows:

  • All populations are initially alive.

  • A round consists of subrounds. In a subround, each alive population is observed once, with the successful ones continuing to the next subround. If there is only one population that is successful in a subround, then if that population currently has a cumulative number of successes that is at least k more than any other population the process stops and that population is declared the best; if not, that population moves to the next subround. If none of the populations in a subround are successful, then the round ends.

  • At the end of a round, any population whose cumulative number of successes is less than that of another by at least k is no longer alive.

Remarks

 

  1. 1.

    Note that the process ends after a subround which had exactly one successful arm, and that arm’s cumulative number of successes is now at least k higher than all other populations and exactly k larger than at least one population.

  2. 2.

    If we had defined a round by saying that each alive population is observed until it had a failure, then when F is the uniform (0, 1) distribution, the expected number of plays until the first population used has a failure is infinite. On the other hand, defining rounds using subrounds results in the mean number of plays being finite. For instance, suppose the probabilities are the values of iid uniform (0, 1) random variables. Let \(N_i\) denote the number of plays in the first round of the arm with \(i^{th}\) largest success probability. Denoting this probability by \(Y_i\), its density is

    $$\begin{aligned} f_{Y_i} (p) = \frac{ n!}{(i-1)! (n-i)!} p^{n-i} (1-p)^{i-1} dp\,, \quad 0<p<1 \end{aligned}$$

    it follows that \(E[N_i] \le E[ \frac{1}{1 - Y_i} ] < \infty \) when \(i > 1.\) In addition,

    $$\begin{aligned} N_1 \le k + \max _{2 \le i \le n} N_i \le k + \sum _{i=2}^n N_i \end{aligned}$$

    giving that

    $$\begin{aligned} E[N_1] \le k + \sum _{i=2}^n E[N_i] < \infty \end{aligned}$$
  3. 3.

    The PW rule as defined in [15] and [16] was such that the populations are initially randomly ordered. In each round, the alive populations were observed in that order, with each population being observed until it had a failure. If at any time one of the populations had k fewer successes than another population, then the former is no longer alive. The process ends when only a single population is alive, and that population is declared best. Thus, for instance, in the original version if the first population observed has k successes in a row, then that population is declared best.

4.1 Analysis of PW

To begin, suppose there are only 2 arms and that their success probabilities are \(p_1 > p_2.\) Suppose we are going to choose an arm by using the procedure which in each round plays each arm until it has a failure, and then stopping at the end of a round if one of the arms has had a cumulative number of successes that is at least k more than the other, with that arm then being chosen. Let \(q_i = 1-p_i, i=1,2,\) and let \(\,X_{i,r} ,\, i = 1,2, \, r \ge 1,\) be independent with \(\; P(X_{i,r} = j) = q_i\,p_i^j, j \ge 0.\) Interpret \(X_{i,r}\) as the number of successes of arm i in round r,  and let \(Y_r = X_{1,r} - X_{2,r},\, r \ge 1.\) Then,

$$\begin{aligned} E[e^{\theta Y_r}] = \frac{q_1}{ 1 - p_1 e^{\theta } } \frac{q_2}{ 1 - p_2 e^{\theta } } \end{aligned}$$

It is now easy to check that \(E[e^{\theta Y_r}] = 1\) if \(e^{\theta } = p_2/p_1\). That is, \( E[ (p_2/p_1)^{Y_r}] = 1. \) If we now let \(\,S_m = \sum _{i=1}^m Y_i,\,\) then \(\; (p_2/p_1)^{S_m}, \, m \ge 1\, \) is a martingale with mean 1. Letting

$$\begin{aligned} N = \min (m: S_m \ge k \;\;\text{ or } \;\; S_m \le -k ) \end{aligned}$$

it follows by the martingale stopping theorem (see [11]) that

$$\begin{aligned} E[ (p_2/p_1)^{S_N}] = 1 \end{aligned}$$

Let \(\,p = P(S_N \ge k)\,\) be the probability that arm 1 is chosen. Then,

$$\begin{aligned} 1= & {} E[ (p_2/p_1)^{S_N}] \\= & {} E[ (p_2/p_1)^{S_N} | S_N \ge k] p + E[ (p_2/p_1)^{S_N} | S_N \le -k] (1-p) \end{aligned}$$

Letting \(X_i, i = 1,2,\) have the distribution of \(X_{i,r},\) it follows, by the lack of memory property of \(X_i\), that

$$\begin{aligned} E[ (p_2/p_1)^{S_N} | S_N \ge k]= & {} (p_2/p_1)^k E[(p_2/p_1)^{X_1}] = (p_2/p_1)^k(q_1/q_2) \\ E[ (p_2/p_1)^{S_N} | S_N \le - k]= & {} (p_2/p_1)^{-k} E[(p_2/p_1)^{-X_2}] = (p_1/p_2)^k (q_2/q_1) \end{aligned}$$

Substituting back yields that

$$\begin{aligned} p = \frac{ 1 - (p_1/p_2)^k (q_2/q_1) }{ (p_2/p_1)^k(q_1/q_2) - (p_1/p_2)^k (q_2/q_1) } \end{aligned}$$
(5)

Conditioning on which arm wins yields that

$$\begin{aligned} E[S_N] = (k + E[X_1] )p + (-k - E[X_2])(1-p) \end{aligned}$$

Letting \(\, m_i = E[X_i] = 1/q_i - 1 = p_i/q_i,\) the preceding gives

$$\begin{aligned} E[S_N] = p( m_1 + m_2 + 2k) - (m_2 + k) \end{aligned}$$

Wald’s equation yields

$$\begin{aligned} E[N]= & {} \frac{ p( m_1 + m_2 + 2k) - m_2 - k }{m_1 - m_2} \end{aligned}$$
(6)

Because \(X_{1,r} + X_{2,r} + 2 \) is the number of plays in round r,  it follows that the total number of plays, call it T, is \(\sum _{r=1}^N (X_{1,r} + X_{2,r} + 2).\) Applying Wald’s equation and using (6) gives

$$\begin{aligned} E[T] = ( p( m_1 + m_2 + 2k) - m_2 - k ) \; \frac{ m_1 + m_2 + 2}{m_1 - m_2} \end{aligned}$$
(7)

Now, suppose that we utilize PW. Let \(B(p_1, p_2)\) and \(M(p_1, p_2)\) be, respectively, the probability that the arm with value \(p_1\) is chosen and the mean number of plays before stopping. From Eq. (5), we have

$$\begin{aligned} B(p_1, p_2) = \frac{ 1 - (p_1/p_2)^k (q_2/q_1) }{ (p_2/p_1)^k(q_1/q_2) - (p_1/p_2)^k (q_2/q_1) } \end{aligned}$$
(8)

Because PW would stop play once the winning arm is ahead by k, whereas E[T] is the mean number of plays when we continue on until a failure occurs, we obtain by conditioning on which arm wins that

$$\begin{aligned} M(p_1, p_2)= & {} E[T] - \frac{ B(p_1, p_2)}{q_1} - \frac{1 - B(p_1, p_2) }{q_2 } \nonumber \\= & {} \left( B(p_1, p_2) ( m_1 + m_2 + 2k) - m_2 - k \right) \; \frac{ m_1 + m_2 + 2}{m_1 - m_2} \nonumber \\{} & {} - \frac{ B(p_1, p_2)}{q_1} - \frac{1 - B(p_1, p_2) }{q_2 } \end{aligned}$$
(9)

where \(m_i = p_i/q_i.\)

Now, suppose there are n arms whose means are the values of independent random variables with distribution F, and let C be the event that the PW policy chooses the best arm. As in our analysis of VT, suppose that all arms participate in each round. Let arm 0 be the best arm, and randomly number the other arms as \(1, \ldots , n-1.\) Say that the best arm beats arm i if the end of round difference between the cumulative number of successes of arm 0 and arm i is at least k before it is less than or equal to \(-k.\) Letting \(B_i, i = 1, \ldots , n-1,\) be the event that arm 0 beats arm i, we have, by the same arguments as in Sect. 1, the following.

Lemma 4

With \(B \equiv B_1B_2 \cdots B_{n-1} \),

$$\begin{aligned} P(C) \ge P(B) \ge (P(B_1))^{n-1} \end{aligned}$$

Also, \(\,P(B) \le P(B^*),\, \) where \(B^*\) is the event that 0 beats the best of arms \(1, \ldots , n-1.\)

Our preceding analysis yields the following corollary.

Corollary 1

With U and V being independent uniform (0, 1) random variables, and

$$\begin{aligned} X = F^{-1}( U^{1/n} )\,, \;\; Y = F^{-1}( U^{1/n} V) \,, \;\; W = F^{-1}( U^{1/n} V^{1/(n-1)}) \end{aligned}$$
$$\begin{aligned} P(B_1)= & {} E[B( X, Y) ] \\ P(B^*)= & {} E[ B(X, W) ] \end{aligned}$$

Also, if we let M denote the mean number of plays, then

$$\begin{aligned} M \approx A = (n-1) E[M(X,Y)] + E[M(X,W)] \end{aligned}$$

4.2 PW with Early Elimination

Suppose we use PW with critical number k and add an early elimination on any population whose first j observations are all failures. Let \(B_e\) be the event that the best population is eliminated early. Letting \(f(p) = F'(p)\) be the density function of the prior distribution, the success probability of the best population has density function \(f_b(p) = n F^{n-1}(p) f(p), \,0< p < 1.\) Consequently, it follows that

$$\begin{aligned} P(B_e) = \int _0 ^1 (1-p)^j n F^{n-1}(p) f(p)dp \end{aligned}$$

Let \(N_{nb}\) be the number of nonbest populations that are eliminated early. To compute \(E[N_{nb}],\) note that the probability a randomly chosen population is eliminated early is \( \int _0^1 (1-p)^j f(p) dp,\) giving that

$$\begin{aligned} n \int _0^1 (1-p)^j f(p) dp = E[\text{ number } \text{ eliminated } \text{ early}] = E[N_{nb}] + P(B_e) \end{aligned}$$

Hence,

$$\begin{aligned} E[N_{nb}] = n \int _0^1 (1-p)^j f(p) dp - \int _0 ^1 (1-p)^j n F^{n-1}(p) f(p)dp \end{aligned}$$

When F is the uniform (0, 1) distribution

$$\begin{aligned} P(B_e) = \int _0 ^1 (1-p)^j n p^{n-1} dp = \frac{ n! j!}{(n+j)!} \end{aligned}$$

and

$$\begin{aligned} E[N_{nb}] = \frac{n}{j+1} - \frac{ n! j!}{(n+j)!} \end{aligned}$$

For instance, if F is the uniform (0, 1) distribution, then when \(n=10, j=5\), we have \(P(B_e) = 0.000333\) and \(E[N_{nb}]= 1.666.\)

4.3 VT Versus PW

Based on numerical experiments, VT and PW have roughly similar performances when \(F(x) = x. \)

Example 4

When \(n=5,\) simulation yielded the following results for PW.

figure b

Thus, choosing PW with \(k=42\) with probability .25 and \(k=43\) with probability .75 results in \(P(C) = .950,\) and requires, on average, 325.795 observations, which is slightly less than the average of 328.56 which, as shown in Example 3, can be obtained by a randomization of VT rules to obtain \(P(C) = .95.\) On the other hand, if we wanted \(\alpha = .954,\) then both VT with \(k=10\) and PW with \(k=48\) achieve that, with VT having a mean of 358.4 observations, compared to 375.4 for PW. (Because the average number of trials needed for PW with \(k=47\) is 367.05,  randomizing between PW(47) and PW(48) still would not be as good as VT(10).)

5 The Normal Case

5.1 VT Rule in the Normal Case

Suppose that observations on population i are the values of independent normal random variables with mean \(\mu _i\) and variance \(\sigma ^2, i = 1, \ldots , n,\) where \(\sigma ^2\) is known and \(\mu _1, \ldots , \mu _n\) are the unknown values of n independent standard normal random variables. As before our objective is to determine, using relatively few observations, the population \(i^*\) such that \(i^* = \text{ argmax }\, \mu _i\) under the proviso that the probability of a correct decision is at least some specified value \(\alpha .\) The VT rule with parameter \(c > 0\) is as follows:

  • Initially, all populations are alive

  • A round consists of a single observation from each alive population.

  • At the end of a round, a population is no longer alive if its cumulative sum of observed values is more than c less than that of another population. (That is, if A is the set of alive populations after round \(k-1\), and \(S_i(k)\) is the cumulative sum of the first k observations of population i, then \(i \in A\) would no longer be alive after round k if \(S_i(k) < \max _{j \in A} S_j(k) - c.\))

  • If only one population is alive, stop and declare it best. Otherwise perform a new round.

Before showing how to determine the appropriate value of c, we present some preliminaries concerning normal partial sums.

5.2 Some Preliminaries

Let \(\Phi \) be the standard normal distribution function. Define

$$\begin{aligned} R(a) = \frac{ \Phi (a)}{ 1 - \Phi (a) } \end{aligned}$$
(10)

Lemma 5

If W is a normal random variable with mean \(\mu \) and variance 1, then

$$\begin{aligned} E[e^{- 2 \mu W}|W> 0]= & {} R(-\mu ) \\ E[W|W > 0]= & {} \mu + \frac{ e^{ -\mu ^2/2} }{\sqrt{ 2 \pi } \Phi (\mu )} \\ E[e^{- 2 \mu W}|W< 0]= & {} R(\mu ) \\ E[W|W < 0]= & {} \mu - \frac{ e^{ -\mu ^2/2} }{\sqrt{ 2 \pi } (1 - \Phi (\mu ) } \end{aligned}$$

Proof

 

$$\begin{aligned} E[e^{- 2 \mu W}|W> 0]= & {} \frac{1}{ \sqrt{ 2 \pi } P(W > 0)} \int _0 ^{\infty } e^{- 2 \mu x} e^{- (x - \mu )^2/2} dx \\= & {} \frac{1}{ \sqrt{ 2 \pi } \Phi (\mu )} \int _0 ^{\infty } e^{-(x+\mu )^2/2} dx \\= & {} \frac{ 1-\Phi (\mu )}{ \Phi (\mu )} \end{aligned}$$

Let \(Z = W- \mu \).

$$\begin{aligned} E[W|W> 0]= & {} \mu + E[Z| Z > - \mu ] \\= & {} \mu + \frac{1}{ \sqrt{ 2 \pi } \Phi (\mu )} \int _{ - \mu }^{\infty } x e^{- x^2/2} \, dx \\= & {} \mu + \frac{e^{- \mu ^2/2}}{ \sqrt{ 2 \pi } \Phi (\mu ) } \end{aligned}$$

Because \(E[e^{-2 \mu W}] = 1,\) the third equality follows from the first upon using the identity

$$\begin{aligned} 1 = E[e^{-2 \mu W}|W > 0] \Phi (\mu ) + E[e^{-2 \mu W}|W < 0] (1 -\Phi (\mu ) ) \end{aligned}$$

Similarly, the fourth equality follows from the second since \(\; \mu = E[W|W > 0] \Phi (\mu ) + E[W|W < 0] (1 -\Phi (\mu ) ). \) \(\square \)

Lemma 6

Let \(S_n = \sum _{i=1}^n W_i,\; n \ge 1,\) where \(W_i, i \ge 1 \) are independent normal random variables with mean \(\mu > 0\) and variance 1. For given \(b > 0,\) let \(N = \min \{n: \;\text{ either } \; S_n < - b \;\text{ or }\; S_n > b\}.\)

  1. (a)

    \( \; R(-\mu ) e^{- 2 \mu b}< E[e^{- 2 \mu S_N}| S_N > b] < e^{- 2 \mu b} \)

  2. (b)

    \( e^{2 \mu b}< E[e^{- 2 \mu S_N}| S_N< - b] < e^{2 \mu b} R(\mu ) \)

Proof

The right hand inequality of (a) is immediate since \(\mu > 0.\) To prove the left side of (a), note that conditional on \(S_N > b\) and on the value \(S_{N-1},\) that \(S_N\) is distributed as b plus the amount by which a normal with mean \(\mu \) and variance 1 exceeds the positive amount \(b - S_{N-1}\) given that it does exceed that amount. But a normal conditioned to be positive is known to have strict increasing failure rate (see [2]), implying that \(S_N | \{S_N > b, S_{N-1} \}\) is stochastically smaller than \( b + W_i| \{W_i > 0\}\). As this is true no matter what the value of \(S_{N-1},\) it follows that \(S_N | \{S_N > b \}\) is stochastically smaller than \( b + W_i| \{W_i > 0\}\), implying that \(E[e^{- 2 \mu S_N}| S_N> b]> e^{-2 \mu b} E[e^{- 2 \mu W_i}|W_i > 0] .\) The result now follows from Lemma 5.

The left hand inequality of (b) is immediate. To prove the right hand inequality, note that the same argument as used in part (a) shows that \(S_N | \{S_N< - b\} >_{st} -b + W_i | \{W_i < 0\} ,\) implying that \(E[e^{- 2 \mu S_N}| S_N< -b ]< e^{2 \mu b} E[e^{- 2 \mu W_i}|W_i < 0] .\) Thus, the result follows from Lemma 5. \(\square \)

Proposition 2

Let \(S_n = \sum _{i=1}^n W_i,\; n \ge 1,\) where \(W_i, i \ge 1 \) are independent normal random variables with mean \(\mu > 0\) and variance 1. For given \(b > 0,\) let \(N = \min \{n: \;\text{ either } \; S_n < - b \;\text{ or }\; S_n > b\}.\)

$$\begin{aligned} \frac{ e^{2 \mu b} - 1}{ e^{2 \mu b} - R(-\mu ) e^{- 2 \mu b} }< P(S_N > b) < \frac{ e^{2 \mu b} R(\mu ) - 1}{ e^{2 \mu b} R(\mu ) - e^{-2 \mu b} } \end{aligned}$$
(11)

Proof

Let \(p = P(S_N > b).\) Because \(E[e^{-2 \mu W_i}] = 1,\) it follows that \(\{e^{- 2 \mu S_n}, n \ge 1\}\) is a martingale with mean 1. Hence, by the martingale stopping theorem

$$\begin{aligned} 1= & {} E[e^{- 2 \mu S_N}] \\= & {} E[e^{- 2 \mu S_N}| S_N > b] p + E[e^{- 2 \mu S_N}| S_N < - b] (1-p) \end{aligned}$$

Hence,

$$\begin{aligned} p = \frac{ E[e^{- 2 \mu S_N}| S_N< - b] - 1}{ E[e^{- 2 \mu S_N}| S_N < - b] - E[e^{- 2 \mu S_N}| S_N > b] } \end{aligned}$$
(12)

Because \(\;\frac{ x - 1}{x-y},\, 0<y<1<x,\,\) increases in both x and y, the inequalities (11) now follow from Lemma 6. \(\square \)

Although we do not directly use the following proposition, it is of independent interest.

Proposition 3

With N as previously defined.

$$\begin{aligned} E[N]\le & {} \frac{ e^{2 \mu b} R(\mu ) - 1}{ e^{2 \mu b} R(\mu ) - e^{-2 \mu b} } \, \left( \frac{2b}{\mu } + \frac{ e^{- \mu ^2/2} }{\mu \sqrt{ 2 \pi } \Phi ( \mu ) } + 1\right) - \frac{b}{\mu } \end{aligned}$$
(13)
$$\begin{aligned} E[N]\ge & {} \frac{ e^{2 \mu b} - 1}{ e^{2 \mu b} - R(-\mu ) e^{- 2 \mu b} } \, \left( \frac{2b}{\mu } - 1 + \frac{ e^{ -\mu ^2/2} }{ \mu \sqrt{ 2 \pi } \left( 1 - \Phi (\mu ) \right) } \right) - \frac{b}{\mu } \nonumber \\{} & {} + \;1 - \frac{ e^{ -\mu ^2/2} }{\mu \sqrt{ 2 \pi } (1 - \Phi (\mu ) )} \end{aligned}$$
(14)

Proof

Wald’s equation gives

$$\begin{aligned} E[N] \mu= & {} E[S_N|S_N> b] p + E[S_N |S_N < - b] (1-p) \\\le & {} (b + E[W|W > 0] ) p -b (1-p) \\= & {} p\left( 2b + \frac{ e^{- \mu ^2/2} }{ \sqrt{ 2 \pi } \Phi ( \mu ) } + \mu \right) - b \end{aligned}$$

where the first inequality used, as shown in Lemma 6, that \(S_N | \{S_N > b \}\) is stochastically smaller than \( b + W| \{W> 0\}\). Inequality (13) now follows from Proposition 2. The lower bound follows from

$$\begin{aligned} E[N] \mu= & {} E[S_N|S_N > b] p + E[S_N |S_N< - b] (1-p) \\\ge & {} b p + (-b + E[W_1|W_1 < 0] )(1-p) \\= & {} \left( 2b - \mu + \frac{ e^{ -\mu ^2/2} }{\sqrt{ 2 \pi } (1 - \Phi (\mu ) } \right) p -b + \mu - \frac{ e^{ -\mu ^2/2} }{\sqrt{ 2 \pi } (1 - \Phi (\mu ) } \end{aligned}$$

where the inequality used that \(S_N | \{S_N < - b \}\) is stochastically larger than \( -b + W_i | \{W_i< 0\}.\) Inequality (12) now follows from Proposition 2. \(\square \)

Remark

One way to approximate \(p = P(S_N > b)\) and E[N] is to “neglect the excess” and assume \(S_N|S_N > b \approx _{st} b\) and \(S_N|S_N < -b] \approx _{st} -b.\) From (12), this gives that

$$\begin{aligned} p \approx \frac{ e^{ 2 \mu b} -1}{ e^{ 2 \mu b} - e^{ -2 \mu b} } \end{aligned}$$
(15)

Also, \(\; \mu E[N] \approx b p - b(1-p) ,\;\) and so (15) gives that

$$\begin{aligned} E[N] \approx \frac{ 2b( e^{ 2 \mu b} -1)}{\mu ( e^{ 2 \mu b} - e^{ -2 \mu b})} - \frac{b}{\mu } \end{aligned}$$
(16)

Example 5

If \(b = 3, \mu = 1,\) then (13), (14), and (16) yield that

$$\begin{aligned} 2.9838 \le E[N] \le 4.2842 , \quad E[N] \approx 2.9852 \end{aligned}$$

Corollary 2

Let \(S_n = \sum _{i=1}^n V_i,\; n \ge 1,\) where \(V_i, i \ge 1 \) are independent normal random variables with mean \(\mu > 0\) and variance \(2 \sigma ^2.\) For given \(c > 0,\) let \(N_{\sigma } = \min \{n: \;\text{ either } \; S_n < - c \;\text{ or }\; S_n > c\}.\)

$$\begin{aligned} \frac{ e^{ \mu c/\sigma ^2} - 1}{ e^{ \mu c/\sigma ^2} - R\left( - \frac{ \mu }{ \sigma \sqrt{2}} \right) e^{- \mu c/\sigma ^2} }< P(S_{N_{\sigma } } > c) < \frac{ e^{ \mu c/\sigma ^2} R\left( \frac{ \mu }{ \sigma \sqrt{2}} \right) - 1}{ e^{ \mu c/\sigma ^2} R\left( \frac{ \mu }{ \sigma \sqrt{2}} \right) - e^{- \mu c/\sigma ^2} } \end{aligned}$$

Moreover,

$$\begin{aligned} E[N_{\sigma } ] \approx \frac{ 2c \, ( e^{ \mu c/\sigma ^2} -1)}{\mu ( e^{ \mu c/\sigma ^2} - e^{ - \mu c/\sigma ^2})} - \frac{c}{\mu } \end{aligned}$$

Proof

Let \(W_i = \frac{V_i}{\sigma \sqrt{2}},\) note that \(E[W_i] = \frac{ \mu }{\sigma \sqrt{2}}.\) Now, using \(b = \frac{ c}{\sigma \sqrt{2}},\) apply Proposition 2 and Eq. (16). \(\square \)

5.3 Analyzing the VT Rule in the Normal Case

We can obtain a lower bound and an effective approximation to P(C),  the probability that the correct choice is made, by a similar argument as in the Bernoulli case. Letting population 0 be the one with the largest mean, and randomly numbering the others as \(1, \ldots , n-1,\) we imagine a “gambler’s ruin" game between populations 0 and i in which the winner is the first one whose cumulative sum is at least c more than that of the other. With \(B_i\) being the event that the best wins this game against population i, and B the event that the best wins all these games, we can show exactly as before that

$$\begin{aligned} P(C) \ge P(B) \ge (P(B_1))^{n-1}, \quad P(B) \le P(B^*) \end{aligned}$$

where \(P(B^*) \) is the probability that the best beats the population with second largest mean. Given the values \(\mu _0, \mu _1, \ldots , \mu _{n-1}\), the difference between the value of a population 0 observation and one from a different population is a normal random variable with variance \(2 \sigma ^2.\) Letting

$$\begin{aligned} L(\mu ) = \frac{ e^{ \mu c/\sigma ^2} - 1}{ e^{ \mu c/\sigma ^2} - R( - \frac{ \mu }{ \sigma \sqrt{2}} ) e^{- \mu c/\sigma ^2} } , \quad U(\mu ) = \frac{ e^{ \mu c/\sigma ^2} R( \frac{ \mu }{ \sigma \sqrt{2}} ) - 1}{ e^{ \mu c/\sigma ^2} R( \frac{ \mu }{ \sigma \sqrt{2}} ) - e^{- \mu c/\sigma ^2} } \end{aligned}$$
(17)

be the lower and upper bounds on \(P(S_{N_{\sigma } } > c) ,\) Corollary 2 yields the following proposition.

Proposition 4

Let U and V be independent uniform (0, 1) random variables, and let

$$\begin{aligned} X = \Phi ^{-1}( U^{1/n} ) - \Phi ^{-1}( U^{1/n} V) ,\qquad Y = \Phi ^{-1}( U^{1/n} ) - \Phi ^{-1}( U^{1/n} V^{1/(n-1) } ) \end{aligned}$$

Then,

$$\begin{aligned} P(C) \ge P(B) \ge (E[ L(X)] )^{n-1} \quad \text{ and } \quad P(B) \le E[ U(Y) ] \end{aligned}$$

Letting N be the number of observations, we can approximate E[N] by approximating the mean number of plays of each of the non-best populations by the mean number of plays in their game against the best population, and approximating the mean number of plays of the best population by the mean number of plays in its game against the second best one. Hence, using Eq. (16) we see that

$$\begin{aligned} E[N] \approx A \equiv (n-1) E[M(X)] + E[M(Y)] \end{aligned}$$
(18)

where

$$\begin{aligned} M(\mu ) = \frac{ 2c \, ( e^{ \mu c/\sigma ^2} -1)}{\mu ( e^{ \mu c/\sigma ^2} - e^{ - \mu c/\sigma ^2})} - \frac{c}{\mu } \end{aligned}$$

The following table compares the performance of VT with the most quoted algorithms of the recent literature: LIL-UCB, TrackAndStop, and Chernoff. The LIL-UCB algorithm (see [10]) uses upper confidence bounds (UCB) based on the law of the iterated logarithm for the expected reward of the arms. At each stage, it uses the arm with the largest upper bound, similar to the UCB algorithm of bandit problems. We use a heuristic variation of the LIL-UCB which has been shown to perform somewhat better than the original [10]. The TrackAndStop algorithm in [7] tracks lower bounds on the optimal proportions of the arm draws and uses a stopping rule based on Chernoff’s work on Generalized Likelihood Ratio statistic. The Chernoff algorithm is similar to TrackAndStop, but rather than track the optimal proportions it instead chooses between the empirical best and second-best. In each of 10, 000 simulation runs, we generate the n means by generating n standard normals and then, simulate the results of the different algorithms (Table 2).

Table 2 \(\bar{N} \) is the average number of plays in 10,000 simulation runs

6 Conclusions

We have presented a Bayesian model for finding the population having the largest mean when the populations under consideration are either all Bernoulli or all normal with a fixed variance. In both cases, we take a Bayesian approach that assumes the unknown means are the values of independent random variables having a specified distribution F. We consider two old rules that had previously been analyzed under the assumption that the largest mean differs from the second largest mean by at least some known positive number. The first of these rules is the vector at a time rule (VT) which in each round takes a sample from each population, eliminates any population whose cumulative sum after a round is at least k less than that of another population, and continues until one population is left. The second old rule, applicable in the Bernoulli case, is play the winner rule which in each, but the last round continues to sample from each remaining population until it has a failure. For a given constant k, we present easily computed bounds and approximations of the probability these rules yield the correct choice and the mean number of observations that are required. We also present numerical evidence showing in the Bernoulli case that the VT rule resulting when F is the uniform (0, 1) distribution has good results, when compared with more recent algorithms that make no assumptions about the set of means, even when the set of means does not look like it came from a uniform (0, 1) distribution. Although we recommend in any problem instance that one utilizes one’s prior knowledge to determine the appropriate prior F, it is comforting to know that the method appears to work well even in cases when the actual means do not appear to have come from F.