Paper Reading AI Learner

Improving MAE against CCE under Label Noise

2019-03-28 17:27:05
Xinshao Wang, Elyor Kodirov, Yang Hua, Neil M. Robertson

Abstract

Label noise is inherent in many deep learning tasks when the training set becomes large. A typical approach to tackle noisy labels is using robust loss functions. Categorical cross entropy (CCE) is a successful loss function in many applications. However, CCE is also notorious for fitting samples with corrupted labels easily. In contrast, mean absolute error (MAE) is noise-tolerant theoretically, but it generally works much worse than CCE in practice. In this work, we have three main points. First, to explain why MAE generally performs much worse than CCE, we introduce a new understanding of them fundamentally by exposing their intrinsic sample weighting schemes from the perspective of every sample's gradient magnitude with respect to logit vector. Consequently, we find that MAE's differentiation degree over training examples is too small so that informative ones cannot contribute enough against the non-informative during training. Therefore, MAE generally underfits training data when noise rate is high. Second, based on our finding, we propose an improved MAE (IMAE), which inherits MAE's good noise-robustness. Moreover, the differentiation degree over training data points is controllable so that IMAE addresses the underfitting problem of MAE. Third, the effectiveness of IMAE against CCE and MAE is evaluated empirically with extensive experiments, which focus on image classification under synthetic corrupted labels and video retrieval under real noisy labels.

Abstract (translated)

当训练集变大时,标签噪声是许多深层学习任务固有的。处理噪声标签的典型方法是使用鲁棒损失函数。分类交叉熵(CCE)在许多应用中是一种成功的损失函数。然而,CCE也因容易贴上损坏标签的样品而臭名昭著。相比之下,平均绝对误差(MAE)在理论上具有抗噪声能力,但在实际应用中通常比CCE效果差得多。在这项工作中,我们有三个要点。首先,为了解释MAE通常比CCE表现差得多的原因,我们从每个样本相对于logit向量的梯度幅度的角度,从根本上揭示了它们的内在样本加权方案,从而引入了对它们的新理解。因此,我们发现MAE对培训实例的区分程度太小,因此在培训过程中,信息型的MAE不能对非信息型的MAE做出足够的贡献。因此,当噪声率较高时,MAE通常会在其训练数据下面。其次,基于我们的发现,我们提出了一种改进的MAE(IMAE),它继承了MAE良好的噪声鲁棒性。此外,训练数据点之间的差异程度是可控的,因此IMAE可以解决MAE的欠匹配问题。第三,通过大量的实验,对IMAE对CCE和MAE的有效性进行了实证评估,重点研究了合成污损标签下的图像分类和真实噪声标签下的视频检索。

URL

https://arxiv.org/abs/1903.12141

PDF

https://arxiv.org/pdf/1903.12141.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot