MarkTechPost@AI 2024年08月20日
mhGPT: Advancing Mental Health AI with a Lightweight, Expert Knowledge-Infused Transformer for Low-Resource Environments
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

mhGPT 是一款轻量级的生成式预训练Transformer模型,专为心理健康文本分析而设计。该模型在心理健康相关的社交媒体和PubMed文章上进行训练,并通过微调在五项特定任务中表现优于MentaLLaMA等最先进的模型,尽管其参数和训练数据更少。mhGPT 的关键创新包括使用专家知识注入的数据、自定义分词器以及NEFTune,以提高其在不平衡数据集上的性能。这项研究表明,mhGPT 有潜力通过优化更小的模型架构来增强心理健康AI,特别是在资源有限的环境中。

🧠 **专家知识注入的数据:** mhGPT 的训练数据包括心理健康相关的社交媒体和PubMed文章,这些数据包含了丰富的专家知识,使模型能够更好地理解和分析心理健康相关文本。

🚀 **轻量级架构:** 与其他大型语言模型相比,mhGPT 的参数更少,但其性能却毫不逊色。这使得 mhGPT 能够在资源有限的设备上运行,并降低了训练和部署的成本。

💪 **定制分词器:** mhGPT 使用了一个定制的分词器,能够更好地处理心理健康文本中的特殊词汇和短语,提高了模型的准确性和效率。

📈 **NEFTune 优化:** NEFTune 是一种用于缓解过拟合的微调技术,特别是在不平衡数据集上。mhGPT 通过使用 NEFTune,能够在有限的数据上取得更好的性能。

💡 **性能优势:** 在多项心理健康文本分析任务中,mhGPT 的表现优于 MentaLLaMA、MentalBERT 和 MentalRoBERTa 等其他模型,证明了其在该领域的有效性和潜力。

🌐 **应用前景:** mhGPT 有望在心理健康领域得到广泛应用,例如:检测心理疾病症状、评估心理健康状况、提供心理健康咨询等。

🙌 **未来发展:** 研究人员将继续改进 mhGPT,并探索其在其他领域的应用,例如:心理健康教育、心理健康服务评估等。

🧠 **性能提升:** mhGPT 在五项特定任务中表现优于其他模型,包括命名实体识别、分类和问答。

🌐 **广泛应用:** mhGPT 可以应用于各种心理健康相关场景,例如:心理健康筛查、心理健康评估和心理健康咨询。

💡 **未来展望:** 研究人员将继续改进 mhGPT,并探索其在其他领域的应用,例如:心理健康教育、心理健康服务评估等。

Mental health profoundly impacts individuals’ quality of life, yet accessing mental health services can be challenging due to stigma, insufficient workforce, and fragmented care systems. NLP has demonstrated its potential in this area, with models developed to detect symptoms and evaluate depression from clinical texts. Language models like BERT have also been adapted for classifying mental disorders. However, creating these models requires substantial computational power, which many organizations need more, and regulations such as HIPAA and GDPR further complicate using cloud-based resources.

Children’s National Hospital and George Washington University researchers introduced mhGPT, a lightweight generative model trained on mental health-related social media and PubMed articles. Designed for low-resource environments, mhGPT, with only 1.98 billion parameters, outperformed larger models like MentaLLaMA and Gemma despite using just 5% of the dataset. The model benefits from integrating diverse mental health data and a custom tokenizer, showing that smaller, expert knowledge-infused models can match or exceed the performance of state-of-the-art models in mental health tasks, even with limited computational resources.

Few studies have developed mental health LLMs, primarily training them on social media data. MentaLLaMA, trained on the interpretable mental health instruction dataset, enhances zero/few-shot mental health analysis. MentalBERT focuses on the early detection of mental disorders and suicidal ideation from social content, outperforming general language models in this domain. Additionally, fine-tuned BERT models on EHR data for specific mental disorders, showing the benefits of domain-specific knowledge transfer. Fine-tuning remains essential for improving LLM performance, with methods like LoRA and QLoRA enabling efficient fine-tuning in low-resource environments by reducing memory usage and training time.

The study utilized 49,812 PubMed Central articles on mental health and over 1 million Reddit submissions and comments from various mental health subreddits. The data was preprocessed by removing irrelevant content and then sampled using two methods: truncating to 512 tokens or chunking with a sliding window. The training involved three configurations using the GPT-NeoX architecture with different parameter sizes and tokenizers. The models were trained on high-performance computing clusters and Amazon EC2 instances. Fine-tuning employed LoRA and QLoRA techniques, with NEFTune applied to mitigate overfitting, particularly in imbalanced datasets.

The study found that mhGPT outperformed comparable models like MentaLLaMA, MentalBERT, and MentalRoBERTa on various tasks despite these models being trained on larger datasets. mhGPT outperformed human annotators in a Named Entity Recognition (NER) task. The baseline model, Gemma-2B, performed well in binary and multi-label classification but may lack interpretability in mental health contexts. NEFTune improved fine-tuning on small, imbalanced datasets, allowing mhGPT to surpass larger models like MentaLLaMA-7B. Models A and B also showed strong performance in specific classification tasks.

In conclusion, mhGPT is a compact generative pre-trained transformer designed for mental health text analysis. Trained on mental health-related social media and PubMed articles, mhGPT was fine-tuned on five specific tasks and outperformed state-of-the-art models like MentaLLaMA despite having fewer parameters and training data. Key innovations include using expert knowledge-infused data, a custom tokenizer, and NEFTune for improved performance on imbalanced datasets. The study demonstrates mhGPT’s potential to enhance mental health AI, especially in low-resource settings, by optimizing a smaller model architecture.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 48k+ ML SubReddit

Find Upcoming AI Webinars here

The post mhGPT: Advancing Mental Health AI with a Lightweight, Expert Knowledge-Infused Transformer for Low-Resource Environments appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

mhGPT 心理健康 AI Transformer 轻量级模型
相关文章