While normalizing flows for continuous data have been extensively researched, flows for discrete data have only recently been explored. These prior models, however, suffer from limitations that are distinct from those of continuous flows. Most notably, discrete flow-based models cannot be straightforwardly optimized with conventional deep learning methods because gradients of discrete functions are undefined or zero. Previous works approximate pseudo-gradients of the discrete functions but do not solve the problem on a fundamental level. In addition to that, backpropagation can be computationally burdensome compared to alternative discrete algorithms such as decision tree algorithms. Our approach seeks to reduce computational burden and remove the need for pseudo-gradients by developing a discrete flow based on decision trees -- building upon the success of efficient tree-based methods for classification and regression for discrete data. We first define a tree-structured permutation (TSP) that compactly encodes a permutation of discrete data where the inverse is easy to compute; thus, we can efficiently compute the density value and sample new data. We then propose a decision tree algorithm to build TSPs that learns the tree structure and permutations at each node via novel criteria. We empirically demonstrate the feasibility of our method on multiple datasets.
翻译:虽然对连续数据的正常流进行了广泛的研究,但直到最近才对离散数据的流进行了探索。这些先前的模型存在不同于连续流的局限性。最明显的是,离散流模型不能以传统的深层学习方法直接优化,因为离散函数的梯度没有定义或零。以前的工作是离散函数的近似假梯位,但在基本水平上无法解决问题。此外,与决策树算法等替代离散算法相比,背向反演算法可能是一种繁琐的计算方法。我们的方法是设法减少计算负担,通过开发基于决定树的离散流来消除对假梯位的需要 -- -- 建立基于高效树基方法对离散数据进行分类和回归的成功。我们首先定义了树结构调整的缩略式(TSP),将离散数据编码成易进行反调的数据;因此,我们可以高效率地计算密度值和抽样新数据。我们然后提出一个决定树算法,以建立基于决策树树树的离散流,从而通过新的标准来学习我们每个树结构的多种经验。