Abstract
von Neumann [(1951). Various techniques used in connection with random digits. National Bureau of Standards Applied Math Series 12: 36–38] introduced a simple algorithm for generating independent unbiased random bits by tossing a (possibly) biased coin with unknown bias. While his algorithm fails to attain the entropy bound, Peres [(1992). Iterating von Neumann's procedure for extracting random bits. The Annals of Statistics 20(1): 590–597] showed that the entropy bound can be attained asymptotically by iterating von Neumann's algorithm. Let
$b(n,p)$
denote the expected number of unbiased bits generated when Peres’ algorithm is applied to an input sequence consisting of the outcomes of
$n$
tosses of the coin with bias
$p$
. With
$p=1/2$
, the coin is unbiased and the input sequence consists of
$n$
unbiased bits, so that
$n-b(n,1/2)$
may be referred to as the cost incurred by Peres’ algorithm when not knowing
$p=1/2$
. We show that
$\lim _{n\to \infty }\log [n-b(n,1/2)]/\log n =\theta =\log [(1+\sqrt {5})/2]$
(where
$\log$
is the logarithm to base
$2$
), which together with limited numerical results suggests that
$n-b(n,1/2)$
may be a regularly varying sequence of index
$\theta$
. (A positive sequence
$\{L(n)\}$
is said to be regularly varying of index
$\theta$
if
$\lim _{n\to \infty }L(\lfloor \lambda n\rfloor )/L(n)=\lambda ^\theta$
for all
$\lambda > 0$
, where
$\lfloor x\rfloor$
denotes the largest integer not exceeding
$x$
.) Some open problems on the asymptotic behavior of
$nh(p)-b(n,p)$
are briefly discussed where
$h(p)=-p\log p- (1-p)\log (1-p)$
denotes the Shannon entropy of a random bit with bias
$p$
.
Publisher
Cambridge University Press (CUP)
Subject
Industrial and Manufacturing Engineering,Management Science and Operations Research,Statistics, Probability and Uncertainty,Statistics and Probability