We consider the fast solution of non-smooth optimization problems as resulting for example from the approximation of elliptic free boundary problems of obstacle or Stefan type. Combining well-known concepts of successive subspace correction methods with convex analysis, we derive a new class of multigrid methods which are globally convergent and have logarithmic bounds of the asymptotic convergence rates. The theoretical considerations are illustrated by numerical experiments.