The best neural architecture for a given machine learning problem depends on many factors: not only the complexity and structure of the dataset, but also on resource constraints including latency, compute, energy consumption, etc. Neural architecture search (NAS) for tabular datasets is an important but under-explored problem. Previous NAS algorithms designed for image search spaces incorporate resource constraints directly into the reinforcement learning (RL) rewards. However, for NAS on tabular datasets, this protocol often discovers suboptimal architectures. This paper develops TabNAS, a new and more effective approach to handle resource constraints in tabular NAS using an RL controller motivated by the idea of rejection sampling. TabNAS immediately discards any architecture that violates the resource constraints without training or learning from that architecture. TabNAS uses a Monte-Carlo-based correction to the RL policy gradient update to account for this extra filtering step. Results on several tabular datasets demonstrate the superiority of TabNAS over previous reward-shaping methods: it finds better models that obey the constraints.
翻译:特定机器学习问题的最佳神经结构取决于许多因素:不仅数据集的复杂性和结构,而且还取决于资源限制,包括隐蔽性、计算、能源消耗等。 神经结构搜索(NAS)对于表格数据集来说是一个重要但探索不足的问题。 先前为图像搜索空间设计的NAS算法将资源限制直接纳入了强化学习(RL)奖励。 然而,对于表格数据集中的NAS来说,这一协议常常发现次优性结构。 本文开发了TabNAS, 这是一种新的和更有效的方法, 利用拒绝抽样理念驱动的 RL 控制器处理列表NAS的资源限制。 TabNAS 立即抛弃了任何违反资源限制而没有培训或学习该结构的架构。 TabNAS 使用基于Monte-Carlo的 RL 政策梯度更新来计算额外过滤步骤。 几个表格数据集的结果显示TabNAS 优于先前的奖分制方法:它找到更好的模型来遵守这些制约。