With the rising complexity of numerous novel applications that serve our modern society comes the strong need to design efficient computing platforms. Designing efficient hardware is, however, a complex multi-objective problem that deals with multiple parameters and their interactions. Given that there are a large number of parameters and objectives involved in hardware design, synthesizing all possible combinations is not a feasible method to find the optimal solution. One promising approach to tackle this problem is statistical modeling of a desired hardware performance. Here, we propose a model-based active learning approach to solve this problem. Our proposed method uses Bayesian models to characterize various aspects of hardware performance. We also use transfer learning and Gaussian regression bootstrapping techniques in conjunction with active learning to create more accurate models. Our proposed statistical modeling method provides hardware models that are sufficiently accurate to perform design space exploration as well as performance prediction simultaneously. We use our proposed method to perform design space exploration and performance prediction for various hardware setups, such as micro-architecture design and OpenCL kernels for FPGA targets. Our experiments show that the number of samples required to create performance models significantly reduces while maintaining the predictive power of our proposed statistical models. For instance, in our performance prediction setting, the proposed method needs 65% fewer samples to create the model, and in the design space exploration setting, our proposed method can find the best parameter settings by exploring less than 50 samples.
翻译:随着为现代社会服务的众多新型应用程序越来越复杂,设计高效的计算平台的需求正在增长。但是,设计高效的硬件是一个复杂的多目标问题,涉及多个参数及其相互作用。由于硬件设计涉及大量参数和目标,综合所有可能的组合并不是找到最优解的可行方法。解决这个问题的一个有前途的方法是统计建模实现所需硬件性能。本文提出了一种基于模型的主动学习方法来解决这个问题。我们提出的方法使用贝叶斯模型来表征所需硬件性能的各个方面。我们还结合了迁移学习和高斯回归引导技术以及主动学习来创建更准确的模型。我们提出的统计建模方法提供了足够准确的硬件模型,可以同时执行设计空间探索和性能预测。 我们使用我们提出的方法来执行各种硬件设置的设计空间探索和性能预测,例如微结构设计和用于FPGA目标的OpenCL内核。我们的实验表明,我们提出的统计模型所需的样本数量显著减少,同时保持预测能力。例如,在我们的性能预测设置中,所提出的方法只需65%的样本即可创建模型,在设计空间探索设置中,我们提出的方法只需探索不到50个样本即可找到最佳参数设置。