Paper Reading AI Learner

Domain-Specific Improvement on Psychotherapy Chatbot Using Assistant

2024-04-24 19:30:18
Cheng Kang, Daniel Novak, Katerina Urbanova, Yuqing Cheng, Yong Hu

Abstract

Large language models (LLMs) have demonstrated impressive generalization capabilities on specific tasks with human-written instruction data. However, the limited quantity, diversity, and professional expertise of such instruction data raise concerns about the performance of LLMs in psychotherapy tasks when provided with domain-specific instructions. To address this, we firstly propose Domain-Specific Assistant Instructions based on AlexanderStreet therapy, and secondly, we use an adaption fine-tuning method and retrieval augmented generation method to improve pre-trained LLMs. Through quantitative evaluation of linguistic quality using automatic and human evaluation, we observe that pre-trained LLMs on Psychotherapy Assistant Instructions outperform state-of-the-art LLMs response baselines. Our Assistant-Instruction approach offers a half-annotation method to align pre-trained LLMs with instructions and provide pre-trained LLMs with more psychotherapy knowledge.

Abstract (translated)

大语言模型(LLMs)已经在特定任务上展示了令人印象深刻的泛化能力,这些任务使用人类编写的指令数据。然而,这种指令数据的数量有限,多样性较小,专业能力有限,这使得当LLMs获得特定领域的指导时,在心理治疗任务上的表现引起了人们的担忧。为解决这个问题,我们首先提出了基于AlexanderStreet治疗的领域特定辅助指令,然后使用自监督和人工评估来改进预训练LLMs。通过使用自动和人工评估对语言质量进行定量评估,我们观察到,使用心理治疗助手指令预训练的LLMs超越了最先进的LLMs响应基线。我们的辅助指令方法提供了一种半注释方法,使预训练的LLM与指令对齐,并为预训练的LLM提供更多的心理治疗知识。

URL

https://arxiv.org/abs/2404.16160

PDF

https://arxiv.org/pdf/2404.16160.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot