Convolutional neural networks (CNN) are widely used in resource-constrained devices in IoT applications. In order to reduce the computational complexity and memory footprint, the resource-constrained devices use fixed-point representation. This representation consumes less area and energy in hardware with similar classification accuracy compared to the floating-point ones. However, to employ the low-precision fixed-point representation, various considerations to gain high accuracy are required. Although many quantization and re-training techniques are proposed to improve the inference accuracy, these approaches are time-consuming and require access to the entire dataset. This paper investigates the effect of different fixed-point hardware units on CNN inference accuracy. To this end, we provide a framework called Fixflow to evaluate the effect of fixed-point computations performed at hardware level on CNN classification accuracy. We can employ different fixed-point considerations at the hardware accelerators.This includes rounding methods and adjusting the precision of the fixed-point operation's result. Fixflow can determine the impact of employing different arithmetic units (such as truncated multipliers) on CNN classification accuracy. Moreover, we evaluate the energy and area consumption of these units in hardware accelerators. We perform experiments on two common MNIST and CIFAR-10 datasets. Our results show that employing different methods at the hardware level specially with low-precision, can significantly change the classification accuracy.
翻译:为了减少计算复杂性和记忆足迹,受资源限制的装置使用固定点表示法。这种表示法比浮动点表示法消耗的硬硬件面积和能量较少,其分类精确度与浮动点表示法相似。然而,为了使用低精度定点表示法,需要采用各种考虑来提高精确度。虽然提出了许多量化和再培训技术来提高定点操作的准确性,但这些方法很费时,需要查阅整个数据集。本文调查了不同固定点硬件单位对CNN精确度的影响。为此,我们提供了一个称为固定点流的框架,以评价在固定点计算法对CNN分类精确度的硬件水平的影响。我们可以在硬件加速器中采用不同的固定点考虑法。这包括四舍五入法和调整定点操作结果的精确性。固定点计算法可以确定使用不同计算单位(如分解的指数级)对CNNCR的精确度的影响。此外,我们用两组的硬点计算法测试了我们共同的硬度数据区域。我们用这些硬度区域来评估常规的硬度的硬度。