Implicit Neural Representations (INRs) have emerged and shown their benefits over discrete representations in recent years. However, fitting an INR to the given observations usually requires optimization with gradient descent from scratch, which is inefficient and does not generalize well with sparse observations. To address this problem, most of the prior works train a hypernetwork that generates a single vector to modulate the INR weights, where the single vector becomes an information bottleneck that limits the reconstruction precision of the output INR. Recent work shows that the whole set of weights in INR can be precisely inferred without the single-vector bottleneck by gradient-based meta-learning. Motivated by a generalized formulation of gradient-based meta-learning, we propose a formulation that uses Transformers as hypernetworks for INRs, where it can directly build the whole set of INR weights with Transformers specialized as set-to-set mapping. We demonstrate the effectiveness of our method for building INRs in different tasks and domains, including 2D image regression and view synthesis for 3D objects. Our work draws connections between the Transformer hypernetworks and gradient-based meta-learning algorithms and we provide further analysis for understanding the generated INRs. The project page with code is at \url{https://yinboc.github.io/trans-inr/} .
翻译:近些年来,出现了隐性内隐表(INRs),并展示了在离散表示中的好处。然而,将IRR与特定观测相匹配通常需要从零开始以梯度下降优化,这种下降效率低,而且没有零散观测效果。为解决这一问题,大多数以前的工作都训练了一个产生单一矢量的超网络来调节IRR的重量,使单个矢量成为限制INR产出重建精确度的信息瓶颈。最近的工作表明,在不采用基于梯度的元学习单维卡瓶颈的情况下,可以精确地推断INR的全部重力。我们的工作受到基于梯度的元学习的普遍表述的驱动,我们提议一种配置,将变异器用作IRR的超网络,从而直接用专门设定到定位的变异器来构建整套IRR的重量。我们展示了在不同任务和领域建立IRR的方法的有效性,包括2D图像回归和查看3D对象的合成。我们的工作在变换机/基于梯度的元数据网络和基于梯度的代谢码分析中,我们提供了I-abrbral 和Lisal-dealisalisalisalisals。