There is an increasing interest from ML and HCI communities in empowering creators with better generative models and more intuitive interfaces with which to control them. In music, ML researchers have focused on training models capable of generating pieces with increasing long-range structure and musical coherence, while HCI researchers have separately focused on designing steering interfaces that support user control and ownership. In this study, we investigate through a common framework how developments in both models and user interfaces are important for empowering co-creation where the goal is to create music that communicates particular imagery or ideas (e.g., as is common for other purposeful tasks in music creation like establishing mood or creating accompanying music for another media). Our study is distinguished in that it measures communication through both composer's self-reported experiences, and how listeners evaluate this communication through the music. In an evaluation study with 26 composers creating 100+ pieces of music and listeners providing 1000+ head-to-head comparisons, we find that more expressive models and more steerable interfaces are important and complementary ways to make a difference in composers communicating through music and supporting their creative empowerment.
翻译:ML和HCI社区越来越关注赋予创作者以更好的基因模型和更直觉的界面来控制这些模型。在音乐方面,ML研究人员侧重于能够产生作品的培训模型,这种模型能够随着远距离结构和音乐的一致性而不断增强,而HCI研究人员则分别侧重于设计支持用户控制和拥有的指导界面。在这项研究中,我们通过一个共同框架,调查模型和用户界面的发展对于赋予创作者权能的重要性,这种互动的目的是创造能够传播特定图像或思想的音乐(例如,对于音乐创作中的其他目的性任务,例如建立情绪或为其他媒体创造音乐是常见的)。我们的研究被区别在于它衡量通过作曲家自我报告的经验进行交流的方式,以及听众如何通过音乐来评价这种交流。在一项评价研究中,26位作曲家制作了100+的音乐和听众提供1000+头对头的比较,我们发现,更清晰的模型和更可导的界面是重要的和互补的方法,通过音乐来改变作曲家的交流方式,支持他们的创造性能力。