We study the meta-learning for support (i.e. the set of non-zero entries) recovery in high-dimensional Principal Component Analysis. We reduce the sufficient sample complexity in a novel task with the information that is learned from auxiliary tasks. We assume each task to be a different random Principal Component (PC) matrix with a possibly different support and that the support union of the PC matrices is small. We then pool the data from all the tasks to execute an improper estimation of a single PC matrix by maximising the $l_1$-regularised predictive covariance to establish that with high probability the true support union can be recovered provided a sufficient number of tasks $m$ and a sufficient number of samples $ O\left(\frac{\log(p)}{m}\right)$ for each task, for $p$-dimensional vectors. Then, for a novel task, we prove that the maximisation of the $l_1$-regularised predictive covariance with the additional constraint that the support is a subset of the estimated support union could reduce the sufficient sample complexity of successful support recovery to $O(\log |J|)$, where $J$ is the support union recovered from the auxiliary tasks. Typically, $|J|$ would be much less than $p$ for sparse matrices. Finally, we demonstrate the validity of our experiments through numerical simulations.
翻译:我们研究用于支持的元学习(即非零条目集)在高层次主元件分析中进行回收(即非零条目集)的元数据分析。我们通过从辅助任务中获取的信息,减少新任务中足够复杂的样本。我们假定每项任务都是一个不同的随机主元元元元元(PC)矩阵,可能有不同的支持,而且个人元矩阵的支持组合规模较小。然后,我们从所有任务中汇集数据,以便通过将1美元固定的预测变量最大化来对单一个人元体矩阵进行不适当的估计,为此将1美元固定的预测变量最大化,从而确定在极有可能情况下,真实支持联盟能够恢复到足够多的任务数量,从每件任务中提供足够数量的样本($(frac\log)\log(p)\ ⁇ m ⁇ right)和足够数量的样本($(fraft) 美元(fraftleft (fraft)\left (frac\\log)\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\