We address the problem of how to achieve optimal inference in distributed quantile regression without stringent scaling conditions. This is challenging due to the non-smooth nature of the quantile regression (QR) loss function, which invalidates the use of existing methodology. The difficulties are resolved through a double-smoothing approach that is applied to the local (at each data source) and global objective functions. Despite the reliance on a delicate combination of local and global smoothing parameters, the quantile regression model is fully parametric, thereby facilitating interpretation. In the low-dimensional regime, we establish a finite-sample theoretical framework for the sequentially defined distributed QR estimators. This reveals a trade-off between the communication cost and statistical error. We further discuss and compare several alternative confidence set constructions, based on inversion of Wald and score-type tests and resampling techniques, detailing an improvement that is effective for more extreme quantile coefficients. In high dimensions, a sparse framework is adopted, where the proposed doubly-smoothed objective function is complemented with an $\ell_1$-penalty. We show that the corresponding distributed penalized QR estimator achieves the global convergence rate after a near-constant number of communication rounds. A thorough simulation study further elucidates our findings.
翻译:我们解决了如何在不严格缩放条件下在分布式四分位回归中实现最佳推断的问题。这具有挑战性,因为四分位回归(QR)损失函数的非光滑性质使得现有方法的使用无效。困难通过对本地(每个数据来源)和全球目标功能采用的双向移动方法来解决。尽管依赖当地和全球平滑参数的微妙组合,但量化回归模型完全对准,从而便利解释。在低维系统中,我们为按顺序定义的分布式QR估计值建立了有限缩放理论框架。这揭示了通信成本和统计错误之间的权衡。我们进一步讨论和比较了基于瓦尔德和得分类型测试的倒置和重新标注技术的若干替代信任套建构,详细说明了对更极端的量化系数有效的改进。在高维度方面,采用了一个稀薄的框架,其中拟议的双向目标功能以1美元/R1美元的比例基调算。我们分配了相应的全球统一率之后,我们进一步展示了相应的折射数。