Feature acquisition algorithms address the problem of acquiring informative features while balancing the costs of acquisition to improve the learning performances of ML models. Previous approaches have focused on calculating the expected utility values of features to determine the acquisition sequences. Other approaches formulated the problem as a Markov Decision Process (MDP) and applied reinforcement learning based algorithms. In comparison to previous approaches, we focus on 1) formulating the feature acquisition problem as a MDP and applying Monte Carlo Tree Search, 2) calculating the intermediary rewards for each acquisition step based on model improvements and acquisition costs and 3) simultaneously optimizing model improvement and acquisition costs with multi-objective Monte Carlo Tree Search. With Proximal Policy Optimization and Deep Q-Network algorithms as benchmark, we show the effectiveness of our proposed approach with experimental study.
翻译:购置特性算法处理获取信息特征的问题,同时平衡购置成本,以提高ML模型的学习绩效。以前的做法侧重于计算各种功能的预期效用值,以确定购置序列。其他方法将这一问题表述为Markov决策程序(MDP)和应用强化学习算法。与以前的做法相比,我们侧重于:(1)将特征获取问题表述为MDP,并应用蒙特卡洛树搜索系统;(2)根据模型改进和购置成本计算每个购置步骤的中间奖励;(3)同时优化模型改进和购置成本,同时以多目标蒙特卡洛树搜索系统为基准。我们以准政策优化和深Q网络算法作为基准,展示了我们提议的实验研究方法的有效性。