Pretraining a neural network on a large dataset is becoming a cornerstone in machine learning that is within the reach of only a few communities with large-resources. We aim at an ambitious goal of democratizing pretraining. Towards that goal, we train and release a single neural network that can predict high quality ImageNet parameters of other neural networks. By using predicted parameters for initialization we are able to boost training of diverse ImageNet models available in PyTorch. When transferred to other datasets, models initialized with predicted parameters also converge faster and reach competitive final performance.
翻译:在大型数据集上对神经网络进行预先培训,正成为机器学习的基石,只有少数资源丰富的社区能够进行这种学习。我们的目标是实现培训前民主化的雄心勃勃的目标。为了实现这一目标,我们培训和发布一个单一的神经网络,可以预测其他神经网络的高质量图像网络参数。通过使用初始化的预测参数,我们能够加强对PyTorch现有多种图像网络模型的培训。在转移到其他数据集后,带有预测参数的模型也会更快地集中并达到有竞争力的最终性能。</s>