ISSN 0439-755X
CN 11-1911/B
主办:中国心理学会
   中国科学院心理研究所
出版:科学出版社

心理学报 ›› 2025, Vol. 57 ›› Issue (6): 987-1000.doi: 10.3724/SP.J.1041.2025.0987 cstr: 32110.14.2025.0987

• 第二十七届中国科协年会学术论文 • 上一篇    下一篇

基于大语言模型的自杀意念文本数据增强与识别技术

章彦博1,2,, 黄峰1,2,3,, 莫柳铃4, 刘晓倩1,2, 朱廷劭1,2()   

  1. 1中国科学院心理研究所行为科学重点实验室, 北京 100101
    2中国科学院大学心理学系, 北京 100049
    3香港城市大学计算学院数据科学系, 香港 999077
    4南开大学社会学院社会心理学系, 天津 300350
  • 收稿日期:2024-02-08 发布日期:2025-04-15 出版日期:2025-06-25
  • 通讯作者: 朱廷劭, E-mail: tszhu@psych.ac.cn
  • 作者简介:

    章彦博和黄峰为共同第一作者。

  • 基金资助:
    国家自然科学基金面上项目(62272206);北京市自然科学基金(IS23088)

Suicidal ideation data augmentation and recognition technology based on large language models

ZHANG Yanbo1,2,, HUANG Feng1,2,3,, MO Liuling4, LIU Xiaoqian1,2, ZHU Tingshao1,2()   

  1. 1CAS Key Laboratory of Behavioral Science, Institute of Psychology, Chinese Academy of Sciences, Beijing 100101, China
    2Department of Psychology, University of Chinese Academy of Sciences, Beijing 100049, China
    3Department of Data Science, College of Computing, City University of Hong Kong, Hong Kong SAR 999077, China
    4Department of Social Psychology, School of Sociology, Nankai University, Tianjin 300350, China
  • Received:2024-02-08 Online:2025-04-15 Published:2025-06-25

摘要:

自杀已成为全球性公共卫生难题, 传统的自杀意念识别方法主要依赖患者主动求助, 而基于文本分析的自动识别模型则受限于标注数据的稀缺性。本研究创新性地提出一种基于大语言模型的数据增强方法, 旨在提升自杀意念文本识别的精度。研究采用双阶段设计:研究1聚焦于数据增强, 研究2验证增强效果。在研究1中, 选用ChatGLM3-6B和Qwen-7B-Chat作为底层模型, 结合有监督学习策略与零样本和少样本学习方法, 优化训练数据集质量。通过8组严谨的对比实验, 结果显示两类自研模型在数据增强方面表现卓越,其处理后数据集的综合得分分别达到0.90和0.92, 显著优于基线模型(p < 0.001)。研究2进一步评估了数据增强对识别模型性能的影响, 结果表明, 增强后的模型在识别准确率和正确拒绝率指标上全面超越最佳基线模型(p < 0.001)。本研究不仅验证了基于大语言模型的数据增强方法在提升自杀意念识别模型性能方面的有效性, 还为心理健康领域的人工智能应用开辟了新方向。这种方法有望在保护用户隐私的同时, 提供及时、有效的自杀风险早期预警, 为自杀预防工作提供重要的技术支持和研究思路。未来研究可着眼于扩大数据异质性、优化提示工程设计、引入人机交互范式等, 进一步拓展该方法在促进临床心理诊断领域的应用。

关键词: 自杀意念, 数据增强, 自杀文本识别, 大语言模型, 人工智能

Abstract:

Suicide constitutes a significant global public health challenge, with the World Health Organization reporting substantial annual mortality rates. Traditional suicide detection methods primarily depend on self-assessment scales and clinical evaluations, which require considerable resources and rely on patients actively seeking assistance. The integrated motivational-volitional (IMV) model offers a theoretical framework for comprehending suicidal behavior progression, with suicidal ideation serving as a critical risk indicator. While text-based analysis presents a promising non-invasive approach for early identification, it encounters technical challenges due to limited annotated data and linguistic complexity. Large Language Models (LLMs) offer unprecedented capabilities in language understanding and generation, potentially addressing these challenges through their ability to comprehend diverse expressions of suicidal ideation and generate high-quality training data.

This research employed a two-stage design leveraging LLMs to address the challenge of limited training data for suicidal ideation recognition. In Study I, we selected ChatGLM3-6B and Qwen-7B-Chat as foundation LLMs and implemented both zero-shot and few-shot learning approaches combined with supervised learning strategies. We extracted examples from an original dataset of Weibo comments to create high-quality training data for the LLMs. Comparative experiments evaluated model performance, with human coders assessing the quality of LLM-generated texts using established suicide risk evaluation criteria. In Study II, we evaluated the impact of LLM-based data augmentation on recognition models by comparing traditional machine learning approaches with LLM-based methods trained on both original and augmented datasets, measuring performance through accuracy and true negative rate metrics.

In Study I, the two self-developed LLM-based models demonstrated excellent performance in suicidal ideation data augmentation, significantly outperforming baseline models according to comprehensive evaluation metrics. The success of these LLM-enhanced models highlighted the effectiveness of high-quality data construction through advanced language modeling capabilities. In Study II, all experimental models trained on LLM-augmented data significantly outperformed their corresponding baseline models in both accuracy and true negative rate. The highest-performing model utilized the ChatGLM3-6B architecture with few-shot learning, showing marked improvements compared to its baseline counterpart. These findings demonstrate the substantial impact of LLM-based data augmentation on model generalization ability, particularly in capturing diverse and subtle expressions of suicidal ideation that traditional approaches often miss.

This study validates the effectiveness of LLM-based data augmentation methods in enhancing suicidal ideation recognition while addressing data scarcity challenges. The non-invasive approach developed through LLM technology has the potential to provide timely and effective early warning of suicide risk while protecting user privacy. This research contributes to both theoretical understanding of LLMs' capabilities in complex psychological text processing and practical applications in mental health monitoring. Future research should explore cross-platform applicability of LLMs, model interpretability, and ethical considerations to further advance this promising technology in suicide prevention and broader mental health applications.

Key words: suicidal ideation, data augmentation, suicide text recognition, large language models, artificial intelligence

中图分类号: