Properties such as composability and automatic differentiation made artificial neural networks a pervasive tool in applications. Tackling more challenging problems caused neural networks to progressively become more complex and thus difficult to define from a mathematical perspective. We present a general definition of linear layer arising from a categorical framework based on the notions of integration theory and parametric spans. This definition generalizes and encompasses classical layers (e.g., dense, convolutional), while guaranteeing existence and computability of the layer's derivatives for backpropagation.
翻译:更具有挑战性的问题使神经网络变得日益复杂,因此难以从数学角度加以界定。我们提出了一个基于集成理论和参数跨度概念的绝对框架产生的线性层的一般定义。这一定义概括并包括了古典层(例如密度大、富集),同时保证了该层衍生物的存在和可乘性,以便进行回馈。