We present a Gaussian Process - Latent Class Choice Model (GP-LCCM) to integrate a non-parametric class of probabilistic machine learning within discrete choice models (DCMs). Gaussian Processes (GPs) are kernel-based algorithms that incorporate expert knowledge by assuming priors over latent functions rather than priors over parameters, which makes them more flexible in addressing nonlinear problems. By integrating a Gaussian Process within a LCCM structure, we aim at improving discrete representations of unobserved heterogeneity. The proposed model would assign individuals probabilistically to behaviorally homogeneous clusters (latent classes) using GPs and simultaneously estimate class-specific choice models by relying on random utility models. Furthermore, we derive and implement an Expectation-Maximization (EM) algorithm to jointly estimate/infer the hyperparameters of the GP kernel function and the class-specific choice parameters by relying on a Laplace approximation and gradient-based numerical optimization methods, respectively. The model is tested on two different mode choice applications and compared against different LCCM benchmarks. Results show that GP-LCCM allows for a more complex and flexible representation of heterogeneity and improves both in-sample fit and out-of-sample predictive power. Moreover, behavioral and economic interpretability is maintained at the class-specific choice model level while local interpretation of the latent classes can still be achieved, although the non-parametric characteristic of GPs lessens the transparency of the model.
翻译:我们提出了一个高斯进程 - 低层类选择模型(GP- LCCM),将非参数性的概率机器学习类别纳入离散的选择模型(DCM)中。高斯进程(GPs)是基于内核的算法,它包含专家知识,它假定先于潜在功能,而不是先于参数,从而使其在解决非线性问题时更具灵活性。我们通过将高斯进程纳入 LCCM 结构,目的是改进未观测到异质的离散表达方式。拟议模式将个人在行为性均匀性分类(低度标准级)中被分配为以GPs和同时估算特定类选择模式模式。此外,我们提出并采用期望-混集法算法,以联合估计/推导出GPs内核功能的超度参数。我们的目标是,将GPs 和具体等级模式选择参数纳入 LCMM 分别以两种不同的模式选择性选择性类别(低度类别),并与不同的 LCCCS- 和 REM 的更精确性标准进行比较。