With the purpose of examining biased updates in variance-reduced stochastic gradient methods, we introduce SVAG, a SAG/SAGA-like method with adjustable bias. SVAG is analyzed in a cocoercive root-finding setting, a setting which yields the same results as in the usual smooth convex optimization setting for the ordinary proximal-gradient method. We show that the same is not true for SVAG when biased updates are used. The step-size requirements for when the operators are gradients are significantly less restrictive compared to when they are not. This highlights the need to not rely solely on cocoercivity when analyzing variance-reduced methods meant for optimization. Our analysis either match or improve on previously known convergence conditions for SAG and SAGA. However, in the biased cases they still do not correspond well with practical experiences and we therefore examine the effect of bias numerically on a set of classification problems. The choice of bias seem to primarily affect the early stages of convergence and in most cases the differences vanish in the later stages of convergence. However, the effect of the bias choice is still significant in a couple of cases.
翻译:为了审查差异减少的悬浮梯度方法中的偏差性更新,我们引入了SVAG(SVAG/SAGA)类似、具有可调整偏差的SVAG(SAG/SAGA)方法。SVAG(SVAG)是用可调整的偏差性分析的根基调查环境分析的,这种环境产生的结果与通常普通准偏差方法的平滑曲线优化环境的结果相同。我们表明,在使用偏差更新时,SVAG(SVAG)的情况不同。当操作者是梯度时,其梯度要求比非梯度限制程度要低得多。这突出表明,在分析旨在优化差异减少的方法时,不能只依赖共焦度。我们的分析要么与SAG和SAGA(SAGA)以前已知的趋同条件相匹配,或者改进。但是,在有偏差的情况下,它们仍然与实际经验不相符,因此我们从数字上研究对一系列分类问题的影响。偏见的选择似乎主要影响趋同的早期阶段,在大多数情况下,在后几个趋同阶段中,但偏见选择的影响仍然很大。