In this paper, we introduce the Hessian-Schatten total-variation (HTV) -- a novel seminorm that quantifies the total "rugosity" of multivariate functions. Our motivation for defining HTV is to assess the complexity of supervised learning schemes. We start by specifying the adequate matrix-valued Banach spaces that are equipped with suitable classes of mixed-norms. We then show that HTV is invariant to rotations, scalings, and translations. Additionally, its minimum value is achieved for linear mappings, supporting the common intuition that linear regression is the least complex learning model. Next, we present closed-form expressions for computing the HTV of two general classes of functions. The first one is the class of Sobolev functions with a certain degree of regularity, for which we show that HTV coincides with the Hessian-Schatten seminorm that is sometimes used as a regularizer for image reconstruction. The second one is the class of continuous and piecewise linear (CPWL) functions. In this case, we show that the HTV reflects the total change in slopes between linear regions that have a common facet. Hence, it can be viewed as a convex relaxation (l1-type) of the number of linear regions (l0-type) of CPWL mappings. Finally, we illustrate the use of our proposed seminorm with some concrete examples.
翻译:在本文中,我们引入了赫森- 夏特总变换( HTV) -- -- 一种新颖的分母,它量化了多变函数的“ 精度” 。 我们定义赫森特的动机是评估受监督的学习计划的复杂性。 我们首先指定了适当的基底价值的巴纳赫空间, 这些空间配备了适当的混合温度类别。 然后我们展示了HTV对旋转、缩放和翻译的不易变性。 此外,它对于线性映射达到了最低值,它支持了普通直觉,即线性回归是最不复杂的学习模式。 其次,我们展示了计算两类功能的HTV的封闭形式表达方式。 第一个是具有某种一定规律性的索博列夫功能的类别。 我们展示了HTV与赫森- 沙特恩半温值相吻合,有时被用作图像重建的正规化因素。 第二个是连续和直线性( CPWL) 的类别。 在此情况下, 我们展示了HTV反映我们平面平面图在直线区域之间最终的变换式。