In this paper, we propose deep partial least squares for the estimation of high-dimensional nonlinear IV regression. As a precursor to a flexible deep neural network architecture, our methodology uses partial least squares (PLS) for dimension reduction and feature selection from the set of instruments and covariates. A central theoretical result, due to Brillinger, shows that the feature selection provided by PLS is consistent and the weights are estimated up to a proportionality constant. We illustrate our methodology with synthetic datasets with a sparse and correlated network structure, together with and draw applications to the effect of childbearing on the mother's labor supply based on classic data of Angrist and Evans (1996). The results on synthetic data as well as applications show that the deep partial least squares method significantly outperforms other related methods. Finally, we conclude with directions for future research.
翻译:在本文中,我们提出用于估计高维非线性四级回归的深度最小方块。作为灵活深神经网络结构的前奏,我们的方法使用部分最小方块进行尺寸削减和从一组仪器和共差中选择特征。由于布里林杰,一个核心理论结果显示,PLS提供的特征选择是一致的,加权估计达到相称性常数。我们用合成数据集展示了我们的方法,其网络结构稀少且相互关联,同时根据安格里斯特和埃文斯的经典数据(1996年),并应用生育对母亲劳动力供应的影响。合成数据和应用结果显示,深度最小方块方法大大优于其他相关方法。最后,我们得出了未来研究的方向。