心理健康
无知
背景(考古学)
任务(项目管理)
水准点(测量)
计算机科学
认知心理学
质量(理念)
人工智能
心理学
数据科学
自然语言处理
精神科
工程类
古生物学
大地测量学
哲学
系统工程
地理
认识论
生物
作者
Kailai Yang,Shaoxiong Ji,Tianlin Zhang,Qianqian Xie,Sophia Ananiadou
出处
期刊:Cornell University - arXiv
日期:2023-01-01
被引量:21
标识
DOI:10.48550/arxiv.2304.03347
摘要
The latest large language models (LLMs) such as ChatGPT, exhibit strong capabilities in automated mental health analysis. However, existing relevant studies bear several limitations, including inadequate evaluations, lack of prompting strategies, and ignorance of exploring LLMs for explainability. To bridge these gaps, we comprehensively evaluate the mental health analysis and emotional reasoning ability of LLMs on 11 datasets across 5 tasks. We explore the effects of different prompting strategies with unsupervised and distantly supervised emotional information. Based on these prompts, we explore LLMs for interpretable mental health analysis by instructing them to generate explanations for each of their decisions. We convey strict human evaluations to assess the quality of the generated explanations, leading to a novel dataset with 163 human-assessed explanations. We benchmark existing automatic evaluation metrics on this dataset to guide future related works. According to the results, ChatGPT shows strong in-context learning ability but still has a significant gap with advanced task-specific methods. Careful prompt engineering with emotional cues and expert-written few-shot examples can also effectively improve performance on mental health analysis. In addition, ChatGPT generates explanations that approach human performance, showing its great potential in explainable mental health analysis.
科研通智能强力驱动
Strongly Powered by AbleSci AI