We study sparse linear regression over a network of agents, modeled as an undirected graph and no server node. The estimation of the $s$-sparse parameter is formulated as a constrained LASSO problem wherein each agent owns a subset of the $N$ total observations. We analyze the convergence rate and statistical guarantees of a distributed projected gradient tracking-based algorithm under high-dimensional scaling, allowing the ambient dimension $d$ to grow with (and possibly exceed) the sample size $N$. Our theory shows that, under standard notions of restricted strong convexity and smoothness of the loss functions, suitable conditions on the network connectivity and algorithm tuning, the distributed algorithm converges globally at a {\it linear} rate to an estimate that is within the centralized {\it statistical precision} of the model, $O(s\log d/N)$. When $s\log d/N=o(1)$, a condition necessary for statistical consistency, an $\varepsilon$-optimal solution is attained after $\mathcal{O}(\kappa \log (1/\varepsilon))$ gradient computations and $O (\kappa/(1-\rho) \log (1/\varepsilon))$ communication rounds, where $\kappa$ is the restricted condition number of the loss function and $\rho$ measures the network connectivity. The computation cost matches that of the centralized projected gradient algorithm despite having data distributed; whereas the communication rounds reduce as the network connectivity improves. Overall, our study reveals interesting connections between statistical efficiency, network connectivity \& topology, and convergence rate in high dimensions.
翻译:我们研究一个代理商网络的细微线性回归,以非方向图形和没有服务器节点的模式建模。对美元偏差参数的估计是一个有限的LASSO问题,其中每个代理商拥有美元总观测的子集。我们分析在高维规模下分布的预测梯度跟踪算法的趋同率和统计保障,使环境维度能够随着(并可能超过)抽样规模的美元增长。我们的理论表明,根据限制的强固凝固和平稳损失功能的标准概念,网络连通和算法调整的合适条件,分布式算法将全球趋同在~线性观察总观察中,每个代理商拥有一个子子子子。我们分析了在高维度规模下分布的梯度追踪算法的趋同率,当$=d/N=o(1),这是统计一致性的一个必要条件时, $lvalepsluslationality-oplationalityalityaility (\\\\\\\\\\\\qlationalliver) laxalyal deal demodeal dealalalalalalalalalations) lax asualtialal 和Orviduview daltialations daltixxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx