When songs are composed or performed, there is often an intent by the singer/songwriter of expressing feelings or emotions through it. For humans, matching the emotiveness in a musical composition or performance with the subjective perception of an audience can be quite challenging. Fortunately, the machine learning approach for this problem is simpler. Usually, it takes a data-set, from which audio features are extracted to present this information to a data-driven model, that will, in turn, train to predict what is the probability that a given song matches a target emotion. In this paper, we studied the most common features and models used in recent publications to tackle this problem, revealing which ones are best suited for recognizing emotion in a cappella songs.
翻译:当歌曲组成或表演时,歌手/作曲家往往有意通过歌曲表达感情或情感。对于人类来说,将音乐成份或表演中的情感与观众的主观感同感相匹配可能是相当具有挑战性的。幸运的是,机器学习这一问题的方法比较简单。通常,它需要一个数据集,从数据集中提取音频功能,将这些信息呈现给数据驱动模型,这反过来又会训练它预测某首歌与目标情感相匹配的可能性。在本文中,我们研究了最近出版物中用来解决这一问题的最常见特征和模型,揭示了哪些最适合在卡普拉歌曲中识别情感。