We study optimization problems where the objective function is modeled through feedforward neural networks with rectified linear unit (ReLU) activation. Recent literature has explored the use of a single neural network to model either uncertain or complex elements within an objective function. However, it is well known that ensembles of neural networks produce more stable predictions and have better generalizability than models with single neural networks, which motivates the investigation of ensembles of neural networks rather than single neural networks in decision-making pipelines. We study how to incorporate a neural network ensemble as the objective function of an optimization model and explore computational approaches for the ensuing problem. We present a mixed-integer linear program based on existing popular big-M formulations for optimizing over a single neural network. We develop a two-phase approach for our model that combines preprocessing procedures to tighten bounds for critical neurons in the neural networks with a Lagrangian relaxation-based branch-and-bound approach. Experimental evaluations of our solution methods suggest that using ensembles of neural networks yields more stable and higher quality solutions, compared to single neural networks, and that our optimization algorithm outperforms (the adaption of) a state-of-the-art approach in terms of computational time and optimality gaps.
翻译:我们研究的是,在目标功能通过矫正线性单元(ReLU)激活的饲料向前神经网络中建模时,我们研究优化问题;最近的一些文献探索了使用单一神经网络在客观功能中建模不确定或复杂的元素;然而,众所周知,神经网络的集合比单一神经网络的模型产生更稳定的预测,比单一神经网络的模型更加普遍,这促使对神经网络的集合进行调查,而不是对决策管道中的单一神经网络进行调查;我们研究如何将神经网络的共性作为优化模型的客观功能,并探索随之产生的问题的计算方法;我们根据现有的流行的大型M型配方提出混合内线性线性方案,以优化单一神经网络;我们为模型制定两阶段办法,将预处理程序与神经网络中的关键神经系统的紧紧界限结合起来,同时采用拉格朗-以放松为基础的分支和封闭式方法;我们对解决方案方法的实验性评估表明,利用神经网络的组合产生更稳定、更高质量的解决方案,比单一神经系统最优化的网络更稳定、最优的计算方法。