In this paper, we introduce the Hessian-Schatten total variation (HTV) -- a novel seminorm that quantifies the total "rugosity" of multivariate functions. Our motivation for defining HTV is to assess the complexity of supervised-learning schemes. We start by specifying the adequate matrix-valued Banach spaces that are equipped with suitable classes of mixed norms. We then show that the HTV is invariant to rotations, scalings, and translations. Additionally, its minimum value is achieved for linear mappings, which supports the common intuition that linear regression is the least complex learning model. Next, we present closed-form expressions of the HTV for two general classes of functions. The first one is the class of Sobolev functions with a certain degree of regularity, for which we show that the HTV coincides with the Hessian-Schatten seminorm that is sometimes used as a regularizer for image reconstruction. The second one is the class of continuous and piecewise-linear (CPWL) functions. In this case, we show that the HTV reflects the total change in slopes between linear regions that have a common facet. Hence, it can be viewed as a convex relaxation (l1-type) of the number of linear regions (l0-type) of CPWL mappings. Finally, we illustrate the use of our proposed seminorm.
翻译:在本文中,我们引入了赫森-夏特总变异(HTV) -- -- 一种新颖的分母,它量化了多变量函数的“精度”总量。我们定义HTV的动机是评估受监督学习计划的复杂程度。我们首先通过指定适当的矩阵价值班纳奇空间,这些空间配备了适当的混合规范类别。我们然后显示,HTV对旋转、缩放和翻译是无差异的。此外,线性绘图的最小值是达到的,它支持了共同直觉,即线性回归是最不复杂的学习模式。接下来,我们展示了两类普通功能的HTV的闭式表达方式。第一个是具有一定规律性的索博利夫功能的类别。我们为此表明,HTV与海珊-沙特南半调调调相吻合,有时用作图像重建的常规。第二个是连续和线性线性(CPWL) 功能的等级。在这种情况下,我们显示,HTV反映的是我们平面平坦型线性区域之间最后的平面图变。