Paper Reading AI Learner

You Only Need Half: Boosting Data Augmentation by Using Partial Content

2024-05-05 06:57:40
Juntao Hu, Yuan Wu

Abstract

We propose a novel data augmentation method termed You Only Need hAlf (YONA), which simplifies the augmentation process. YONA bisects an image, substitutes one half with noise, and applies data augmentation techniques to the remaining half. This method reduces the redundant information in the original image, encourages neural networks to recognize objects from incomplete views, and significantly enhances neural networks' robustness. YONA is distinguished by its properties of parameter-free, straightforward application, enhancing various existing data augmentation strategies, and thereby bolstering neural networks' robustness without additional computational cost. To demonstrate YONA's efficacy, extensive experiments were carried out. These experiments confirm YONA's compatibility with diverse data augmentation methods and neural network architectures, yielding substantial improvements in CIFAR classification tasks, sometimes outperforming conventional image-level data augmentation methods. Furthermore, YONA markedly increases the resilience of neural networks to adversarial attacks. Additional experiments exploring YONA's variants conclusively show that masking half of an image optimizes performance. The code is available at this https URL.

Abstract (translated)

我们提出了一个名为You Only Need hAlf(YONA)的新数据增强方法,该方法简化了数据增强过程。YONA对图像进行分割,用噪声替换图像的一半,并应用数据增强技术对剩余的一半。这种方法减少了原始图像中的冗余信息,鼓励神经网络从 incomplete views 中识别物体,并显著增强了神经网络的鲁棒性。YONA的特点是参数免费、简单易用、增强各种现有的数据增强策略,从而提高了神经网络的鲁棒性,而无需增加额外的计算成本。为了证明YONA的有效性,进行了广泛的实验。这些实验证实了YONA与各种数据增强方法和神经网络架构的兼容性,在CIFAR分类任务中取得了显著的改进,有时甚至超过了传统的图像级数据增强方法。此外,YONA显著增强了神经网络对对抗攻击的鲁棒性。此外,通过探索YONA的变体,实验证明遮盖图像的一半可以优化性能。代码可在此处访问:https://www.thunar.mamail.com/

URL

https://arxiv.org/abs/2405.02830

PDF

https://arxiv.org/pdf/2405.02830.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot