Gaussian graphical regression is a powerful means that regresses the precision matrix of a Gaussian graphical model on covariates, permitting the numbers of the response variables and covariates to far exceed the sample size. Model fitting is typically carried out via separate node-wise lasso regressions, ignoring the network-induced structure among these regressions. Consequently, the error rate is high, especially when the number of nodes is large. We propose a multi-task learning estimator for fitting Gaussian graphical regression models; we design a cross-task group sparsity penalty and a within task element-wise sparsity penalty, which govern the sparsity of active covariates and their effects on the graph, respectively. For computation, we consider an efficient augmented Lagrangian algorithm, which solves subproblems with a semi-smooth Newton method. For theory, we show that the error rate of the multi-task learning based estimates has much improvement over that of the separate node-wise lasso estimates, because the cross-task penalty borrows information across tasks. To address the main challenge that the tasks are entangled in a complicated correlation structure, we establish a new tail probability bound for correlated heavy-tailed (sub-exponential) variables with an arbitrary correlation structure, a useful theoretical result in its own right. Finally, the utility of our method is demonstrated through simulations as well as an application to a gene co-expression network study with brain cancer patients.
翻译:Gaussia 图形回归是一种强大的手段, 它可以让 Gaussia 图形模型的精确矩阵在共变量上倒退, 允许响应变量和共变量的数量远远超过样本大小。 模型安装通常通过不同的节点拉索回归进行, 忽略网络引发的结构。 因此, 错误率是很高的, 特别是在节点数量很大时。 我们为安装 Gaussia 图形回归模型建议了一个多任务学习估计符; 我们设计了一个跨任务组宽度罚款和任务元素宽度罚款, 以调整活动共变量的宽度及其在图形中的效果。 计算时, 我们考虑高效地增强Lagrangia 算法, 以半移动牛顿方法解决子问题。 在理论上, 我们显示多任务学习的错误率比不同的不偏重值估算要高得多, 因为交叉任务在任务元素宽度中, 将跨任务值的任意变量 的网络应用, 将一个重度的直径直径直的直径对等的逻辑结构, 将我们用一个重的直径直方的直方的直方的直方的直方的直方的直方的直方的直方的直方的直方的直方计算 。