We analyze the performance of the least absolute shrinkage and selection operator (Lasso) for the linear model when the number of regressors $N$ grows larger keeping the true support size $d$ finite, i.e., the ultra-sparse case. The result is based on a novel treatment of the non-rigorous replica method in statistical physics, which has been applied only to problem settings where $N$ ,$d$ and the number of observations $M$ tend to infinity at the same rate. Our analysis makes it possible to assess the average performance of Lasso with Gaussian sensing matrices without assumptions on the scaling of $N$ and $M$, the noise distribution, and the profile of the true signal. Under mild conditions on the noise distribution, the analysis also offers a lower bound on the sample complexity necessary for partial and perfect support recovery when $M$ diverges as $M = O(\log N)$. The obtained bound for perfect support recovery is a generalization of that given in previous literature, which only considers the case of Gaussian noise and diverging $d$. Extensive numerical experiments strongly support our analysis.
翻译:我们分析线性模型最不绝对的缩缩和选择操作员(Lasso)的性能,当递减者数目增加时,当递减者数目增加时,则线性模型的收缩和选择操作员(Lasso)的性能(Lasso)的性能(Lasso),因为递减者数目的增加使真正的支助规模(美元和美元)保持在一定的限度(美元),即超粗的个案,结果基于对统计物理学中非硬性复制方法的新式处理,该方法只适用于问题环境,因为当美元与美元相差为美元=O(log N)时,问题往往会以同样的速度得到精确的支持。我们的分析使得能够评估Lasso与高斯的感测矩阵的平均性能(Lasso)的性能,而没有假设美元和美元的规模、噪音分布和真实信号的轮廓。在噪音分布的较轻的条件下,分析也对部分和完美回收所需的抽样复杂性提供了较弱的制约,因为当美元与美元相差为美元时,因为美元是O(\ N)。为了获得完全的支持而必须得到的支持而得到的是对先前文献中给出的概述,它只考虑高西亚噪音和差价分析。</s>