Market sentiment analysis on social media content requires knowledge of both financial markets and social media jargon, which makes it a challenging task for human raters. The resulting lack of high-quality labeled data stands in the way of conventional supervised learning methods. Instead, we approach this problem using semi-supervised learning with a large language model (LLM). Our pipeline generates weak financial sentiment labels for Reddit posts with an LLM and then uses that data to train a small model that can be served in production. We find that prompting the LLM to produce Chain-of-Thought summaries and forcing it through several reasoning paths helps generate more stable and accurate labels, while using a regression loss further improves distillation quality. With only a handful of prompts, the final model performs on par with existing supervised models. Though production applications of our model are limited by ethical considerations, the model's competitive performance points to the great potential of using LLMs for tasks that otherwise require skill-intensive annotation.
翻译:有关社交媒体内容的市场情绪分析既需要金融市场知识,也需要社交媒体术语方面的知识,这使得它成为人类评级者的一项艰巨任务。因此,缺乏高质量的标签数据阻碍了常规监管学习方法的采用。相反,我们用一个大型语言模式(LLM)进行半监督学习来解决这个问题。我们的管道为Reddit 职位和LLM制造了微弱的金融情绪标签,然后用这些数据来培训一个在生产中可以使用的小型模型。我们发现,促使LLM制作“连锁搜索摘要”并用多种推理路径迫使它生成更稳定和准确的标签,同时利用回归损失进一步提高了蒸馏质量。由于只有少量的提示,最终模型与现有的监管模式同步运行。尽管我们模型的生产应用受到伦理考虑的限制,但模型的竞争性表现表明,在否则需要技能密集型说明的任务中使用LMS的巨大潜力。