78
0

Can Large Language Models Identify Implicit Suicidal Ideation? An Empirical Evaluation

Abstract

We present a comprehensive evaluation framework for assessing Large Language Models' (LLMs) capabilities in suicide prevention, focusing on two critical aspects: the Identification of Implicit Suicidal ideation (IIS) and the Provision of Appropriate Supportive responses (PAS). We introduce \ourdata, a novel dataset of 1,308 test cases built upon psychological frameworks including D/S-IAT and Negative Automatic Thinking, alongside real-world scenarios. Through extensive experiments with 8 widely used LLMs under different contextual settings, we find that current models struggle significantly with detecting implicit suicidal ideation and providing appropriate support, highlighting crucial limitations in applying LLMs to mental health contexts. Our findings underscore the need for more sophisticated approaches in developing and evaluating LLMs for sensitive psychological applications.

View on arXiv
@article{li2025_2502.17899,
  title={ Can Large Language Models Identify Implicit Suicidal Ideation? An Empirical Evaluation },
  author={ Tong Li and Shu Yang and Junchao Wu and Jiyao Wei and Lijie Hu and Mengdi Li and Derek F. Wong and Joshua R. Oltmanns and Di Wang },
  journal={arXiv preprint arXiv:2502.17899},
  year={ 2025 }
}
Comments on this paper