In practical federated learning scenarios, the participating devices may have different bitwidths for computation and memory storage by design. However, despite the progress made in device-heterogeneous federated learning scenarios, the heterogeneity in the bitwidth specifications in the hardware has been mostly overlooked. We introduce a pragmatic FL scenario with bitwidth heterogeneity across the participating devices, dubbed as Bitwidth Heterogeneous Federated Learning (BHFL). BHFL brings in a new challenge, that the aggregation of model parameters with different bitwidths could result in severe performance degeneration, especially for high-bitwidth models. To tackle this problem, we propose ProWD framework, which has a trainable weight dequantizer at the central server that progressively reconstructs the low-bitwidth weights into higher bitwidth weights, and finally into full-precision weights. ProWD further selectively aggregates the model parameters to maximize the compatibility across bit-heterogeneous weights. We validate ProWD against relevant FL baselines on the benchmark datasets, using clients with varying bitwidths. Our ProWD largely outperforms the baseline FL algorithms as well as naive approaches (e.g. grouped averaging) under the proposed BHFL scenario.
翻译:在实际的联邦学习情景中,参与装置可能具有不同的比特线,用于按设计计算和存储记忆。然而,尽管在设备偏差的联邦学习情景中取得了进展,硬件中比特规格的异质性大多被忽视。我们引入了一种实用的FL情景,在参与装置中具有比特维特异性,称为Bitwith异质联邦学习(BHFL)。BHFL带来了新的挑战,即将不同比特维特的模型参数组合起来,可能导致严重性能退化,特别是高比特比特模型。为了解决这一问题,我们提议了ProWD框架,在中央服务器上有一个可加培训的重量脱硫器,将低比特重逐步重建为更高的位维特异异异异性重量,最后是全精密重量。ProWD进一步有选择地汇总模型参数,以最大限度地实现位异异比重重量的兼容性。我们根据相关FWDD(ProWD),在相关的FWDF-L类中,以我们的标准级的比值基准模型基线,以我们B-L为基准级标准级的BFFFFFFFFFF格式的客户为基础,对基准进行验证。