Integrating discrete probability distributions and combinatorial optimization problems into neural networks has numerous applications but poses several challenges. We propose Implicit Maximum Likelihood Estimation (I-MLE), a framework for end-to-end learning of models combining discrete exponential family distributions and differentiable neural components. I-MLE is widely applicable: it only requires the ability to compute the most probable states; and does not rely on smooth relaxations. The framework encompasses several approaches, such as perturbation-based implicit differentiation and recent methods to differentiate through black-box combinatorial solvers. We introduce a novel class of noise distributions for approximating marginals via perturb-and-MAP. Moreover, we show that I-MLE simplifies to maximum likelihood estimation when used in some recently studied learning settings that involve combinatorial solvers. Experiments on several datasets suggest that I-MLE is competitive with and often outperforms existing approaches which rely on problem-specific relaxations.
翻译:将离散概率分布和组合优化问题整合到神经网络中有许多应用,但提出了若干挑战。我们提出了隐性最大可能性估计(I-MLE),这是一个将离散指数家庭分布和不同神经元组成结合起来的模型的端到端学习框架。I-MLE广泛适用:它只需要能够计算最可能的情况;不依赖顺畅的放松。这个框架包含若干种方法,例如以扰动为基础的隐含差异和最近通过黑盒组合处理器进行区分的方法。我们引入了一种新颖的噪音分配类别,用于通过perturb-and-MAP来接近边缘。此外,我们表明,I-MLE在某些最近研究的学习环境中使用的涉及组合解析器时,会简化到最大的可能性估计。对若干数据集的实验表明,I-MLE具有竞争力,而且往往超越了依赖问题特定放松的现有方法。