Safe use of Deep Neural Networks (DNNs) requires careful testing. However, deployed models are often trained further to improve in performance. As rigorous testing and evaluation is expensive, triggers are in need to determine the degree of change of a model. In this paper we investigate the weight space of DNN models for structure that can be exploited to that end. Our results show that DNN models evolve on unique, smooth trajectories in weight space which can be used to track DNN training progress. We hypothesize that curvature and smoothness of the trajectories as well as step length along it may contain information on the state of training as well as potential domain shifts. We show that the model trajectories can be separated and the order of checkpoints on the trajectories recovered, which may serve as a first step towards DNN model versioning.
翻译:安全使用深神经网络(DNN)需要仔细测试。但是,部署的模型往往经过进一步的培训才能提高性能。由于严格的测试和评估费用昂贵,因此需要触发来确定模型的改变程度。在本文件中,我们调查DNN模型对于可以用于此目的的结构的重量空间。我们的结果表明,DNN模型在可被用于跟踪DNN培训进展的重量空间中的独特、平稳的轨迹上演进。我们假设轨道的曲缩和平稳以及沿轨道的步长可能包含关于培训状况以及潜在域变换的信息。我们显示,模型轨迹可以分离,轨道上的检查站可以恢复,这可作为向DNNN模型版本迈出的第一步。