Kernel segmentation aims at partitioning a data sequence into several non-overlapping segments that may have nonlinear and complex structures. In general, it is formulated as a discrete optimization problem with combinatorial constraints. A popular algorithm for optimally solving this problem is dynamic programming (DP), which has quadratic computation and memory requirements. Given that sequences in practice are too long, this algorithm is not a practical approach. Although many heuristic algorithms have been proposed to approximate the optimal segmentation, they have no guarantee on the quality of their solutions. In this paper, we take a differentiable approach to alleviate the aforementioned issues. First, we introduce a novel sigmoid-based regularization to smoothly approximate the combinatorial constraints. Combining it with objective of the balanced kernel clustering, we formulate a differentiable model termed Kernel clustering with sigmoid-based regularization (KCSR), where the gradient-based algorithm can be exploited to obtain the optimal segmentation. Second, we develop a stochastic variant of the proposed model. By using the stochastic gradient descent algorithm, which has much lower time and space complexities, for optimization, the second model can perform segmentation on overlong data sequences. Finally, for simultaneously segmenting multiple data sequences, we slightly modify the sigmoid-based regularization to further introduce an extended variant of the proposed model. Through extensive experiments on various types of data sequences performances of our models are evaluated and compared with those of the existing methods. The experimental results validate advantages of the proposed models. Our Matlab source code is available on github.
翻译:内核分割法旨在将数据序列分割成若干非线性和复杂结构的非重叠部分。 一般来说, 它是一个离散的优化问题, 有组合限制 。 优化解决这一问题的流行算法是动态编程( DP), 有二次计算和内存要求 。 鉴于实践中的序列太长, 此算法不是实用的方法 。 虽然许多基于渐变的算法都是为了接近最佳分解, 但对于其解决方案的质量没有保证 。 在本文中, 我们用一种不同的方法来缓解上述问题。 首先, 我们引入一种新型的基于结构的优化优化优化优化优化, 以便顺利地接近组合限制 。 将它与平衡内核组合( DP) 的目标合并起来, 具有四分层计算和内核整合的要求。 我们开发了一种不同的模型, 以梯度为基础来获取最佳分解 。 其次, 我们开发了一个我们拟议模型的随机源代码变异变。 通过使用随机渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐渐变的离位计算模型, 将数据序列的模型在最后进行数据序列分析, 将数据序列的模型进行。 将数据转换为我们现有的模型的模型, 将数据序列的模型, 将数据推延延变。