Paper Reading AI Learner

Activity Recognition From Newborn Resuscitation Videos

2023-03-14 11:04:32
Øyvind Meinich-Bache, Simon Lennart Austnes, Kjersti Engan, Ivar Austvoll, Trygve Eftestøl, Helge Myklebust, Simeon Kusulla, Hussein Kidanto, Hege Ersdal

Abstract

Objective: Birth asphyxia is one of the leading causes of neonatal deaths. A key for survival is performing immediate and continuous quality newborn resuscitation. A dataset of recorded signals during newborn resuscitation, including videos, has been collected in Haydom, Tanzania, and the aim is to analyze the treatment and its effect on the newborn outcome. An important step is to generate timelines of relevant resuscitation activities, including ventilation, stimulation, suction, etc., during the resuscitation episodes. Methods: We propose a two-step deep neural network system, ORAA-net, utilizing low-quality video recordings of resuscitation episodes to do activity recognition during newborn resuscitation. The first step is to detect and track relevant objects using Convolutional Neural Networks (CNN) and post-processing, and the second step is to analyze the proposed activity regions from step 1 to do activity recognition using 3D CNNs. Results: The system recognized the activities newborn uncovered, stimulation, ventilation and suction with a mean precision of 77.67 %, a mean recall of 77,64 %, and a mean accuracy of 92.40 %. Moreover, the accuracy of the estimated number of Health Care Providers (HCPs) present during the resuscitation episodes was 68.32 %. Conclusion: The results indicate that the proposed CNN-based two-step ORAAnet could be used for object detection and activity recognition in noisy low-quality newborn resuscitation videos. Significance: A thorough analysis of the effect the different resuscitation activities have on the newborn outcome could potentially allow us to optimize treatment guidelines, training, debriefing, and local quality improvement in newborn resuscitation.

Abstract (translated)

目标:出生窒息是Neonatal 死亡的主要原因之一。生存的关键是进行即时和持续的高质量新生儿复苏。一个关键步骤是在坦桑尼亚哈杰多收集的新生儿复苏期间记录的信号数据集,包括视频,旨在分析治疗及其对新生儿结果的影响。一个重要步骤是在复苏期间生成相关的复苏活动时间线,包括通气、刺激、吸氧和吸痰等。方法:我们提议使用两个步骤的深度神经网络系统——ORAA-net,利用低质量的复苏期间视频录制进行活动识别。第一个步骤是使用卷积神经网络(CNN)和后处理检测和跟踪相关物体。第二个步骤是使用3DCNNs从第一个步骤中分析提议的活动区域。结果:系统在新生儿复苏期间识别了未暴露的活动、刺激、通气和吸痰,其平均精度为77.67%,平均召回率为77.64%,平均准确度为92.40%。此外,估计在复苏期间存在的医疗服务提供者数量的准确性为68.32%。结论:结果显示,提议的基于CNN的两步ORAA-net可以在嘈杂的低质量新生儿复苏视频中进行物体检测和活动识别。 significance:全面分析不同复苏活动对新生儿结果的影响可能使我们优化治疗指南、培训、汇报和当地质量改进新生儿复苏。

URL

https://arxiv.org/abs/2303.07789

PDF

https://arxiv.org/pdf/2303.07789.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot