The human brain has the ability to carry out new tasks with limited experience. It utilizes prior learning experiences to adapt the solution strategy to new domains. On the other hand, deep neural networks (DNNs) generally need large amounts of data and computational resources for training. However, this requirement is not met in many settings. To address these challenges, we propose the TUTOR DNN synthesis framework. TUTOR targets non-image datasets. It synthesizes accurate DNN models with limited available data, and reduced memory and computational requirements. It consists of three sequential steps: (1) drawing synthetic data from the same probability distribution as the training data and labeling the synthetic data based on a set of rules extracted from the real dataset, (2) use of two training schemes that combine synthetic data and training data to learn DNN weights, and (3) employing a grow-and-prune synthesis paradigm to learn both the weights and the architecture of the DNN to reduce model size while ensuring its accuracy. We show that in comparison with fully-connected DNNs, on an average TUTOR reduces the need for data by 6.0x (geometric mean), improves accuracy by 3.6%, and reduces the number of parameters (floating-point operations) by 4.7x (4.3x) (geometric mean). Thus, TUTOR is a less data-hungry, accurate, and efficient DNN synthesis framework.
翻译:人类大脑有能力在经验有限的情况下执行新的任务。它利用先前的学习经验将解决方案战略调整到新的领域。另一方面,深神经网络通常需要大量的数据和计算资源来进行培训。然而,在许多环境下,这一要求没有得到满足。为了应对这些挑战,我们提议TUTOR DNN综合框架。TUTOR针对非模拟数据集。它综合准确的DNN模型,其可用数据有限,记忆和计算要求减少。它包括三个相继步骤:(1)从培训数据的相同概率分布中提取合成数据,并根据从真实数据集提取的一套规则标出合成数据的标签;(2)使用两个将合成数据和培训数据相结合的培训计划来学习DNN的重量;(3)使用增长和复发综合模型模式模式,以缩小模型的大小,同时确保其准确性。我们表明,与完全连接的DNNNU相比,平均TUR数据分布减少了数据需求6.x(地平比值),将数据精确性运行率降低3.6%(正比值基准值),将数据精确性参数降低3.6%。