A convolutional neural network (ConvNet) is usually trained and then tested using images drawn from the same distribution. To generalize a ConvNet to various tasks often requires a complete training dataset that consists of images drawn from different tasks. In most scenarios, it is nearly impossible to collect every possible representative dataset as a priori. The new data may only become available after the ConvNet is deployed in clinical practice. ConvNet, however, may generate artifacts on out-of-distribution testing samples. In this study, we present Targeted Gradient Descent (TGD), a novel fine-tuning method that can extend a pre-trained network to a new task without revisiting data from the previous task while preserving the knowledge acquired from previous training. To a further extent, the proposed method also enables online learning of patient-specific data. The method is built on the idea of reusing a pre-trained ConvNet's redundant kernels to learn new knowledge. We compare the performance of TGD to several commonly used training approaches on the task of Positron emission tomography (PET) image denoising. Results from clinical images show that TGD generated results on par with training-from-scratch while significantly reducing data preparation and network training time. More importantly, it enables online learning on the testing study to enhance the network's generalization capability in real-world applications.
翻译:为了将ConvNet推广到各种任务,往往需要一个完整的培训数据集,其中包括从不同任务中提取的图像。在多数情况下,几乎不可能将每一个可能的具有代表性的数据集作为先验性收集。新数据只有在ConvNet在临床实践中部署后才能提供。ConvNet可能会在分配之外测试样本中生成文物。在本研究中,我们将目标梯级梯级源(TGD)的性能与一些常用的培训方法进行比较,该方法可以将预先训练过的网络扩大到新的任务,而不必在保留从以往培训中获得的知识的同时重新审视先前任务中的数据。在更大程度上,拟议方法还使得能够将每个有代表性的数据集作为先行收集。该方法的基础是重新使用事先训练过的ConvNet的冗余内核来学习新知识。我们将TGD的性能与几个常用的培训方法进行了比较,这些方法可以将预先训练的网络扩大到新的任务,而无需重新审视先前任务中的数据数据。从临床图像的测试结果可以使TGD产生更多的在线学习能力。