The growing diversity of large language models (LLMs) means users often need to compare and combine outputs from different models to obtain higher-quality or more comprehensive responses. However, switching between separate interfaces and manually integrating outputs is inherently inefficient, leading to a high cognitive burden and fragmented workflows. To address this, we present LLMartini, a novel interactive system that supports seamless comparison, selection, and intuitive cross-model composition tools. The system decomposes responses into semantically aligned segments based on task-specific criteria, automatically merges consensus content, and highlights model differences through color coding while preserving unique contributions. In a user study (N=18), LLMartini significantly outperformed conventional manual methods across all measured metrics, including task completion time, cognitive load, and user satisfaction. Our work highlights the importance of human-centered design in enhancing the efficiency and creativity of multi-LLM interactions and offers practical implications for leveraging the complementary strengths of various language models.
翻译:大型语言模型(LLM)的日益多样化意味着用户经常需要比较并整合不同模型的输出,以获得更高质量或更全面的回答。然而,在独立的界面之间切换并手动整合输出本质上效率低下,会导致较高的认知负担和碎片化的工作流程。为解决这一问题,我们提出了LLMartini——一个支持无缝比较、选择及直观跨模型组合工具的新型交互系统。该系统根据任务特定标准将回答分解为语义对齐的片段,自动合并共识内容,并通过颜色编码突出模型差异,同时保留独特贡献。在一项用户研究(N=18)中,LLMartini在所有测量指标上均显著优于传统手动方法,包括任务完成时间、认知负荷和用户满意度。我们的工作凸显了以人为中心的设计在提升多LLM交互效率与创造力方面的重要性,并为利用各类语言模型的互补优势提供了实践启示。