We review Quasi Maximum Likelihood estimation of factor models for high-dimensional panels of time series. We consider two cases: (1) estimation when no dynamic model for the factors is specified (Bai and Li, 2016); (2) estimation based on the Kalman smoother and the Expectation Maximization algorithm thus allowing to model explicitly the factor dynamics (Doz et al., 2012). Our interest is in approximate factor models, i.e., when we allow for the idiosyncratic components to be mildly cross-sectionally, as well as serially, correlated. Although such setting apparently makes estimation harder, we show, in fact, that factor models do not suffer of the curse of dimensionality problem, but instead they enjoy a blessing of dimensionality property. In particular, we show that if the cross-sectional dimension of the data, $N$, grows to infinity, then: (i) identification of the model is still possible, (ii) the mis-specification error due to the use of an exact factor model log-likelihood vanishes. Moreover, if we let also the sample size, $T$, grow to infinity, we can also consistently estimate all parameters of the model and make inference. The same is true for estimation of the latent factors which can be carried out by weighted least-squares, linear projection, or Kalman filtering/smoothing. We also compare the approaches presented with: Principal Component analysis and the classical, fixed $N$, exact Maximum Likelihood approach. We conclude with a discussion on efficiency of the considered estimators.
翻译:本文回顾了在高维时间序列面板中对因子模型进行拟最大似然估计的方法。我们考虑两种情况:一种是在没有为因子指定动态模型的情况下进行估计(Bai and Li, 2016);另一种是基于卡尔曼滤波器和期望最大化算法进行估计,从而明确建立因子动态模型(Doz et al., 2012)。我们的兴趣在于近似因子模型,即允许特异组件在交叉面上略微相关,以及串行相关情况。虽然这种情况似乎使估计更难,但事实上,我们证明了因子模型并不遭受维数灾难问题的困扰,而是享受维数有利性质。特别是,我们证明,如果数据的横截面维度 $N$ 增长为无穷大,则:(i)仍然可以对模型进行识别,(ii)由于使用精确因子模型对数似然的误规模错误消失。此外,如果我们还让样本大小 $T$ 增长到无穷大,我们也可以一致地估计模型的所有参数并进行推断。这对于估计隐含因素也是正确的,可以通过加权最小二乘法、线性投影或卡尔曼滤波/平滑计算。我们还将所述方法与主成分分析和经典的、固定 $N$ 的精确最大似然方法进行了比较。最后,我们对所考虑的估计方法的效率进行了讨论。