Stochastic gradient descent (SGD) is a premium optimization method for training neural networks, especially for learning objectively defined labels such as image objects and events. When a neural network is instead faced with subjectively defined labels--such as human demonstrations or annotations--SGD may struggle to explore the deceptive and noisy loss landscapes caused by the inherent bias and subjectivity of humans. While neural networks are often trained via preference learning algorithms in an effort to eliminate such data noise, the de facto training methods rely on gradient descent. Motivated by the lack of empirical studies on the impact of evolutionary search to the training of preference learners, we introduce the RankNEAT algorithm which learns to rank through neuroevolution of augmenting topologies. We test the hypothesis that RankNEAT outperforms traditional gradient-based preference learning within the affective computing domain, in particular predicting annotated player arousal from the game footage of three dissimilar games. RankNEAT yields superior performances compared to the gradient-based preference learner (RankNet) in the majority of experiments since its architecture optimization capacity acts as an efficient feature selection mechanism, thereby, eliminating overfitting. Results suggest that RankNEAT is a viable and highly efficient evolutionary alternative to preference learning.
翻译:在神经网络面临主观定义的标签时,当神经网络面临主观定义的标签时,例如人类演示或说明-SGD可能会努力探索人类固有的偏见和主观性所造成的欺骗性和吵闹的损失景观。神经网络往往通过偏好学习算法接受培训,以消除这些数据噪音,而事实上的培训方法则依靠梯度下降。我们引进了RankNeAT算法,这种算法通过神经变化变迁的神经变幻变变变变的增殖表层等来进行排名。我们测试了一种假设,即RankNeAT在影响性计算机域内超越传统的梯度偏爱学习,特别是预测了三个不同游戏的游戏镜头中带有注解的玩家,从而消除了这些数据的噪音,而事实上的培训方法则依靠梯度下降。由于缺少关于进化搜索对优先学习者培训的影响的经验性研究,我们引入了RankNeAT算法,这种算法通过增强表层结构的神经变异变变变变能力来进行排名。我们测试RangNAT的假设是高效率的变迁机制,从而消除了高性变换率。