Deep Ensembles (DE) are a prominent approach achieving excellent performance on key metrics such as accuracy, calibration, uncertainty estimation, and out-of-distribution detection. However, hardware limitations of real-world systems constrain to smaller ensembles and lower capacity networks, significantly deteriorating their performance and properties. We introduce Packed-Ensembles (PE), a strategy to design and train lightweight structured ensembles by carefully modulating the dimension of their encoding space. We leverage grouped convolutions to parallelize the ensemble into a single common backbone and forward pass to improve training and inference speeds. PE is designed to work under the memory budget of a single standard neural network. Through extensive studies we show that PE faithfully preserve the properties of DE, e.g., diversity, and match their performance in terms of accuracy, calibration, out-of-distribution detection and robustness to distribution shift.
翻译:深度组合(DE)是一个突出的方法,在精确度、校准、不确定性估计和分配范围外检测等关键指标上取得优异性能。然而,现实世界系统的硬件限制限制了小型组合和低容量网络,使其性能和性能大大恶化。我们引入了包装组合(PE)战略,通过仔细调整编码空间的维度来设计和培训轻量结构组合。我们利用组合组合组合将组合组合平行成一个单一的共同主干网,并向前推进改进培训和推断速度。PE设计在单一标准神经网络的记忆预算下工作。我们通过广泛的研究表明,PE忠实地保存了DE的特性,例如多样性,并在精确度、校准、分配范围外检测和对分布转移的稳健性方面匹配其性能。