Real-world machine learning applications may require functions that are fast-to-evaluate and interpretable. In particular, guaranteed monotonicity of the learned function can be critical to user trust. We propose meeting these goals for low-dimensional machine learning problems by learning flexible, monotonic functions using calibrated interpolated look-up tables. We extend the structural risk minimization framework of lattice regression to train monotonic look-up tables by solving a convex problem with appropriate linear inequality constraints. In addition, we propose jointly learning interpretable calibrations of each feature to normalize continuous features and handle categorical or missing data, at the cost of making the objective non-convex. We address large-scale learning through parallelization, mini-batching, and propose random sampling of additive regularizer terms. Case studies with real-world problems with five to sixteen features and thousands to millions of training samples demonstrate the proposed monotonic functions can achieve state-of-the-art accuracy on practical problems while providing greater transparency to users.
翻译:现实世界机器学习应用可能需要快速评估和可解释的功能。特别是,确保学习功能的单一性对于用户信任至关重要。我们提议通过使用校准的相互交错的图表学习灵活、单体功能,从而达到低维机器学习问题的目标。我们扩大拉蒂斯回归的结构性风险最小化框架,通过解决具有适当的线性不平等限制的连接问题,对单体查看表进行培训。此外,我们提议共同学习每种特性的可解释性校准,以使连续特性正常化,处理绝对或缺失的数据,成本是使目标非电解密。我们提出通过平行化、微型打斗和随机抽样抽样选取添加定律术语来解决大规模学习问题。有5至16个特征和数千万个培训样本的真世界问题的案例研究表明,拟议的单体功能可以在提高用户透明度的同时实现对实际问题的最新精确度。