THESIS
2015
ix, 47 pages : illustrations ; 30 cm
Abstract
The alternating direction method of multipliers (ADMM) is an efficient optimization solver
for a wide variety of machine learning models. Recently, stochastic ADMM has been integrated
with variance reduction methods for stochastic gradient, leading to the SAG-ADMM and SDCA-ADMM
algorithms that have fast convergence rates and low iteration complexities. However, their
space requirements can still be high, particularly when used in large multiclass, multilabel or multitask
learning problems. In this thesis, I propose a novel integration of ADMM with the method of
stochastic variance reduced gradient (SVRG). It retains the benefits of SAG-ADMM and SDCA-ADMM,
but is more advantageous in that its storage requirement is very low, even independent of
the sample size n. Experimental res...[
Read more ]
The alternating direction method of multipliers (ADMM) is an efficient optimization solver
for a wide variety of machine learning models. Recently, stochastic ADMM has been integrated
with variance reduction methods for stochastic gradient, leading to the SAG-ADMM and SDCA-ADMM
algorithms that have fast convergence rates and low iteration complexities. However, their
space requirements can still be high, particularly when used in large multiclass, multilabel or multitask
learning problems. In this thesis, I propose a novel integration of ADMM with the method of
stochastic variance reduced gradient (SVRG). It retains the benefits of SAG-ADMM and SDCA-ADMM,
but is more advantageous in that its storage requirement is very low, even independent of
the sample size n. Experimental results demonstrate that it is as fast as SAG-ADMM and SDCA-ADMM,
but can be used on much bigger data sets.
Post a Comment