Automatically generating short summaries from users' online mental health posts could save counselors' reading time and reduce their fatigue so that they can provide timely responses to those seeking help for improving their mental state. Recent Transformers-based summarization models have presented a promising approach to abstractive summarization. They go beyond sentence selection and extractive strategies to deal with more complicated tasks such as novel word generation and sentence paraphrasing. Nonetheless, these models have a prominent shortcoming; their training strategy is not quite efficient, which restricts the model's performance. In this paper, we include a curriculum learning approach to reweigh the training samples, bringing about an efficient learning procedure. We apply our model on extreme summarization dataset of MentSum posts -- a dataset of mental health related posts from Reddit social media. Compared to the state-of-the-art model, our proposed method makes substantial gains in terms of Rouge and Bertscore evaluation metrics, yielding 3.5% (Rouge-1), 10.4% (Rouge-2), and 4.7% (Rouge-L), 1.5% (Bertscore) relative improvements.
翻译:从用户的在线心理健康站上自动生成简短摘要可以节省顾问的阅读时间,减少他们的疲劳,以便他们能够及时回应那些寻求帮助改善其精神状态的人。最近以变异者为基础的总和模型为抽象的概括化提供了很有希望的方法。它们超越了刑罚选择和采掘战略,而是为了处理更复杂的任务,如新颖的生成单词和句引言。尽管如此,这些模型有一个显著的缺点;它们的培训战略效率不高,限制了模型的绩效。在本文中,我们包括了一种课程学习方法,以调整培训样本,从而形成一个高效的学习程序。我们应用了我们关于MentSum员额的极端加对齐数据集的模式,这是来自Redit社会媒体的与心理健康有关的数据集。与最先进的模型相比,我们拟议的方法在红色和贝尔茨核心评价指标方面取得了巨大的收益,产生了3.5%(Roug-1)、10.4%(Rouge-2)和4.7%(Rouge-L)、1.5%(Bertcore)的相对改进。