Paper Reading AI Learner

Explanation based Bias Decoupling Regularization for Natural Language Inference

2024-04-20 14:20:24
Jianxiang Zang, Hui Liu

Abstract

The robustness of Transformer-based Natural Language Inference encoders is frequently compromised as they tend to rely more on dataset biases than on the intended task-relevant features. Recent studies have attempted to mitigate this by reducing the weight of biased samples during the training process. However, these debiasing methods primarily focus on identifying which samples are biased without explicitly determining the biased components within each case. This limitation restricts those methods' capability in out-of-distribution inference. To address this issue, we aim to train models to adopt the logic humans use in explaining causality. We propose a simple, comprehensive, and interpretable method: Explanation based Bias Decoupling Regularization (EBD-Reg). EBD-Reg employs human explanations as criteria, guiding the encoder to establish a tripartite parallel supervision of Distinguishing, Decoupling and Aligning. This method enables encoders to identify and focus on keywords that represent the task-relevant features during inference, while discarding the residual elements acting as biases. Empirical evidence underscores that EBD-Reg effectively guides various Transformer-based encoders to decouple biases through a human-centric lens, significantly surpassing other methods in terms of out-of-distribution inference capabilities.

Abstract (translated)

基于Transformer的自然语言推理模型的稳健性经常受到数据集偏差的影响,而不是依赖于预期的任务相关特征。为了解决这个问题,最近的研究尝试通过在训练过程中减少偏差样本的权重来减轻这种偏差。然而,这些去偏方法主要关注于确定哪些样本存在偏差,而没有明确确定每个案例中的偏差组件。这个限制限制了这些方法在离散分布推理方面的能力。为了应对这个问题,我们旨在训练模型以模仿人类解释因果关系的逻辑。我们提出了一种简单、全面、可解释的方法:解释偏差去耦正则化(EBD-Reg)。EBD-Reg利用人类解释作为标准,指导编码器建立区分、去耦和align的三角关系。这种方法使编码器能够在推理过程中识别和关注代表任务相关特征的关键字,同时丢弃作为偏见的残余元素。实证证据表明,EBD-Reg有效地通过人本主义视角引导各种Transformer编码器通过去偏来解耦偏差,在离散分布推理能力上显著超越其他方法。

URL

https://arxiv.org/abs/2404.13390

PDF

https://arxiv.org/pdf/2404.13390.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot