Stochastic approximation

1

Stochastic approximation methods are a family of iterative methods typically used for root-finding problems or for optimization problems. The recursive update rules of stochastic approximation methods can be used, among other things, for solving linear systems when the collected data is corrupted by noise, or for approximating extreme values of functions which cannot be computed directly, but only estimated via noisy observations. In a nutshell, stochastic approximation algorithms deal with a function of the form which is the expected value of a function depending on a random variable \xi. The goal is to recover properties of such a function f without evaluating it directly. Instead, stochastic approximation algorithms use random samples of to efficiently approximate properties of f such as zeros or extrema. Recently, stochastic approximations have found extensive applications in the fields of statistics and machine learning, especially in settings with big data. These applications range from stochastic optimization methods and algorithms, to online forms of the EM algorithm, reinforcement learning via temporal differences, and deep learning, and others. Stochastic approximation algorithms have also been used in the social sciences to describe collective dynamics: fictitious play in learning theory and consensus algorithms can be studied using their theory. The earliest, and prototypical, algorithms of this kind are the Robbins–Monro and Kiefer–Wolfowitz algorithms introduced respectively in 1951 and 1952.

Robbins–Monro algorithm

The Robbins–Monro algorithm, introduced in 1951 by Herbert Robbins and Sutton Monro, presented a methodology for solving a root finding problem, where the function is represented as an expected value. Assume that we have a function M(\theta), and a constant \alpha, such that the equation has a unique root at \theta^. It is assumed that while we cannot directly observe the function M(\theta), we can instead obtain measurements of the random variable N(\theta) where. The structure of the algorithm is to then generate iterates of the form: Here, is a sequence of positive step sizes. Robbins and Monro proved, Theorem 2 that \theta_n converges in L^2 (and hence also in probability) to \theta^, and Blum later proved the convergence is actually with probability one, provided that: A particular sequence of steps which satisfy these conditions, and was suggested by Robbins–Monro, have the form: a_n=a/n, for a > 0. Other series, such as are possible but in order to average out the noise in N(\theta), the above condition must be met.

Example

Consider the problem of estimating the mean \theta^* of a probability distribution from a stream of independent samples. Let, then the unique solution to is the desired mean \theta^*. The RM algorithm gives usThis is equivalent to stochastic gradient descent with loss function. It is also equivalent to a weighted average:In general, if there exists some function L such that, then the Robbins–Monro algorithm is equivalent to stochastic gradient descent with loss function L(\theta). However, the RM algorithm does not require L to exist in order to converge.

Complexity results

Subsequent developments and Polyak–Ruppert averaging

While the Robbins–Monro algorithm is theoretically able to achieve O(1/n) under the assumption of twice continuous differentiability and strong convexity, it can perform quite poorly upon implementation. This is primarily due to the fact that the algorithm is very sensitive to the choice of the step size sequence, and the supposed asymptotically optimal step size policy can be quite harmful in the beginning. Chung (1954) and Fabian (1968) showed that we would achieve optimal convergence rate with (or ). Lai and Robbins designed adaptive procedures to estimate such that \theta_n has minimal asymptotic variance. However the application of such optimal methods requires much a priori information which is hard to obtain in most situations. To overcome this shortfall, Polyak (1991) and Ruppert (1988) independently developed a new optimal algorithm based on the idea of averaging the trajectories. Polyak and Juditsky also presented a method of accelerating Robbins–Monro for linear and non-linear root-searching problems through the use of longer steps, and averaging of the iterates. The algorithm would have the following structure:The convergence of to the unique root \theta^* relies on the condition that the step sequence {a_n} decreases sufficiently slowly. That is A1) Therefore, the sequence with satisfies this restriction, but \alpha = 1 does not, hence the longer steps. Under the assumptions outlined in the Robbins–Monro algorithm, the resulting modification will result in the same asymptotically optimal convergence rate yet with a more robust step size policy. Prior to this, the idea of using longer steps and averaging the iterates had already been proposed by Nemirovski and Yudin for the cases of solving the stochastic optimization problem with continuous convex objectives and for convex-concave saddle point problems. These algorithms were observed to attain the nonasymptotic rate. A more general result is given in Chapter 11 of Kushner and Yin by defining interpolated time, interpolated process and interpolated normalized process U^n(\cdot) as Let the iterate average be and the associate normalized error to be. With assumption A1) and the following A2) A2) There is a Hurwitz matrix A and a symmetric and positive-definite matrix \Sigma such that converges weakly to U(\cdot), where U(\cdot) is the statisolution to where w(\cdot) is a standard Wiener process. satisfied, and define . Then for each t,

The success of the averaging idea is because of the time scale separation of the original sequence ' and the averaged sequence ', with the time scale of the former one being faster.

Application in stochastic optimization

Suppose we want to solve the following stochastic optimization problemwhere is differentiable and convex, then this problem is equivalent to find the root \theta^* of. Here Q(\theta,X) can be interpreted as some "observed" cost as a function of the chosen \theta and random effects X. In practice, it might be hard to get an analytical form of, Robbins–Monro method manages to generate a sequence to approximate \theta^* if one can generate , in which the conditional expectation of X_n given \theta_n is exactly , i.e. X_n is simulated from a conditional distribution defined by Here is an unbiased estimator of. If X depends on \theta, there is in general no natural way of generating a random outcome that is an unbiased estimator of the gradient. In some special cases when either IPA or likelihood ratio methods are applicable, then one is able to obtain an unbiased gradient estimator. If X is viewed as some "fundamental" underlying random process that is generated independently of \theta, and under some regularization conditions for derivative-integral interchange operations so that, then gives the fundamental gradient unbiased estimate. However, for some applications we have to use finite-difference methods in which has a conditional expectation close to but not exactly equal to it. We then define a recursion analogously to Newton's Method in the deterministic algorithm:

Convergence of the algorithm

The following result gives sufficient conditions on \theta_n for the algorithm to converge: C1) C2) C3) C4) C5) Then \theta_n converges to \theta^* almost surely. Here are some intuitive explanations about these conditions. Suppose is a uniformly bounded random variables. If C2) is not satisfied, i.e., thenis a bounded sequence, so the iteration cannot converge to \theta^* if the initial guess \theta_0 is too far away from \theta^. As for C3) note that if \theta_n converges to \theta^ then so we must have ,and the condition C3) ensures it. A natural choice would be . Condition C5) is a fairly stringent condition on the shape of g(\theta); it gives the search direction of the algorithm.

Example (where the stochastic gradient method is appropriate)

Suppose, where f is differentiable and is a random variable independent of \theta. Then depends on the mean of X, and the stochastic gradient method would be appropriate in this problem. We can choose

Kiefer–Wolfowitz algorithm

The Kiefer–Wolfowitz algorithm was introduced in 1952 by Jacob Wolfowitz and Jack Kiefer, and was motivated by the publication of the Robbins–Monro algorithm. However, the algorithm was presented as a method which would stochastically estimate the maximum of a function. Let M(x) be a function which has a maximum at the point \theta. It is assumed that M(x) is unknown; however, certain observations N(x), where, can be made at any point x. The structure of the algorithm follows a gradient-like method, with the iterates being generated as where N(x_n+c_n) and N(x_n-c_n) are independent. At every step, the gradient of M(x) is approximated akin to a central difference method with h=2c_n. So the sequence {c_n} specifies the sequence of finite difference widths used for the gradient approximation, while the sequence {a_n} specifies a sequence of positive step sizes taken along that direction. Kiefer and Wolfowitz proved that, if M(x) satisfied certain regularity conditions, then x_n will converge to \theta in probability as n\to\infty, and later Blum in 1954 showed x_n converges to \theta almost surely, provided that: A suitable choice of sequences, as recommended by Kiefer and Wolfowitz, would be a_n = 1/n and.

Subsequent developments and important issues

Further developments

An extensive theoretical literature has grown up around these algorithms, concerning conditions for convergence, rates of convergence, multivariate and other generalizations, proper choice of step size, possible noise models, and so on. These methods are also applied in control theory, in which case the unknown function which we wish to optimize or find the zero of may vary in time. In this case, the step size a_n should not converge to zero but should be chosen so as to track the function. , 2nd ed., chapter 3 C. Johan Masreliez and R. Douglas Martin were the first to apply stochastic approximation to robust estimation. The main tool for analyzing stochastic approximations algorithms (including the Robbins–Monro and the Kiefer–Wolfowitz algorithms) is a theorem by Aryeh Dvoretzky published in 1956.

This article is derived from Wikipedia and licensed under CC BY-SA 4.0. View the original article.

Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc.
Bliptext is not affiliated with or endorsed by Wikipedia or the Wikimedia Foundation.

View original