Paper Reading AI Learner

Neural Language Model for Automated Classification of Electronic Medical Records at the Emergency Room. The Significant Benefit of Unsupervised Generative Pre-training

2019-08-30 17:25:06
Binbin Xu, Cédric Gil-Jardiné, Frantz Thiessard, Eric Tellier, Marta Avalos, Emmanuel Lagarde

Abstract

In the context of a project to build a national injury surveillance system based on emergency room (ER) visit reports, it was necessary to develop a coding system capable of classifying the causes of these visits based on the automatic reading of clinical notes written by clinicians. Supervised learning techniques have shown good results but require manual coding of a large number of texts for model training. New levels of performance have been achieved in neural language models (NLM) with the use of the Transformer architecture with an unsupervised generative pre-training step. Our hypothesis is that this latter method significantly reduces the number of annotated samples required. We derived from available diagnostic codes the traumatic/non-traumatic nature of the cause of the ER visit. We then designed a case study to predict from free text clinical notes whether a visit was traumatic or not. We compared two strategies: Strategy A consisted in training the GPT-2 NLM on the training data (with a maximum of 161930 samples) with all labels (trauma/non-trauma) in a single fully-supervised phase. In Strategy B, we split the training data in two parts, 151930 samples without any label for the self-supervised pre-training phase and a much smaller part (up to 10000 samples) for the supervised fine-tuning with labels. In Strategy A, AUC and F1 score reach 0.97 and 0.89 respectively after the processing of 7000 samples. The use of generative pre-training (Strategy B) achieved an AUC of 0.93 and an F1-score of 0.80 after the processing of only 120 samples. The same performance was achieved with only 30 labeled samples processed 3 times (3 epochs of learning). To conclude, it is possible to easily adapt a multi-purpose NLM model such as the GPT-2 to create a powerful tool for classification of free-text notes with the need of a very small number of labeled samples.

Abstract (translated)

URL

https://arxiv.org/abs/1909.01136

PDF

https://arxiv.org/pdf/1909.01136.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot