Paper Reading AI Learner

Improved Hard Example Mining by Discovering Attribute-based Hard Person Identity

2019-05-06 15:38:36
Xiao Wang, Ziliang Chen, Rui Yang, Bin Luo, Jin Tang

Abstract

In this paper, we propose Hard Person Identity Mining (HPIM) that attempts to refine the hard example mining to improve the exploration efficacy in person re-identification. It is motivated by following observation: the more attributes some people share, the more difficult to separate their identities. Based on this observation, we develop HPIM via a transferred attribute describer, a deep multi-attribute classifier trained from the source noisy person attribute datasets. We encode each image into the attribute probabilistic description in the target person re-ID dataset. Afterwards in the attribute code space, we consider each person as a distribution to generate his view-specific attribute codes in different practical scenarios. Hence we estimate the person-specific statistical moments from zeroth to higher order, which are further used to calculate the central moment discrepancies between persons. Such discrepancy is a ground to choose hard identity to organize proper mini-batches, without concerning the person representation changing in metric learning. It presents as a complementary tool of hard example mining, which helps to explore the global instead of the local hard example constraint in the mini-batch built by randomly sampled identities. Extensive experiments on two person re-identification benchmarks validated the effectiveness of our proposed algorithm.

Abstract (translated)

本文提出了硬人识别挖掘(HPIM),试图对硬人识别挖掘进行改进,以提高挖掘效率。这是由以下观察得出的:一些人共享的属性越多,就越难分离他们的身份。基于这一观察,我们通过一个传递属性描述器(transferred attribute describer)开发了hpim,这是一个从噪声源人属性数据集中训练出来的深度多属性分类器。我们将每个图像编码为目标人物重新识别数据集中的属性概率描述。然后在属性代码空间中,我们将每个人作为一个分布,在不同的实际场景中生成特定于视图的属性代码。因此,我们估计了从零到高阶的特定于人的统计时刻,这进一步用于计算人与人之间的中心时刻差异。这种差异是在不考虑度量学习中人的表征变化的情况下,选择难识别的对象来组织合适的小批量的基础。它作为硬实例挖掘的一个补充工具,有助于探索由随机抽样标识生成的小批量中的全局而不是局部硬实例约束。对两人重新识别基准的大量实验验证了该算法的有效性。

URL

https://arxiv.org/abs/1905.02102

PDF

https://arxiv.org/pdf/1905.02102.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot