Abstract

A new concept of (normalized) convergence of random variables is introduced. This convergence is preserved under Lipschitz transformations, follows from convergence in mean and itself implies convergence in probability. If a sequence of random variables satisfies a limit theorem then it is a normalized convergent sequence. The introduced concept is applied to the convergence rate study of a statistical approach in stochastic optimization.