Paper Reading AI Learner

A Dual Reinforcement Learning Framework for Unsupervised Text Style Transfer

2019-05-24 07:02:05
Fuli Luo, Peng Li, Jie Zhou, Pengcheng Yang, Baobao Chang, Zhifang Sui, Xu Sun

Abstract

Unsupervised text style transfer aims to transfer the underlying style of text but keep its main content unchanged without parallel data. Most existing methods typically follow two steps: first separating the content from the original style, and then fusing the content with the desired style. However, the separation in the first step is challenging because the content and style interact in subtle ways in natural language. Therefore, in this paper, we propose a dual reinforcement learning framework to directly transfer the style of the text via a one-step mapping model, without any separation of content and style. Specifically, we consider the learning of the source-to-target and target-to-source mappings as a dual task, and two rewards are designed based on such a dual structure to reflect the style accuracy and content preservation, respectively. In this way, the two one-step mapping models can be trained via reinforcement learning, without any use of parallel data. Automatic evaluations show that our model outperforms the state-of-the-art systems by a large margin, especially with more than 8 BLEU points improvement averaged on two benchmark datasets. Human evaluations also validate the effectiveness of our model in terms of style accuracy, content preservation and fluency. Our code and data, including outputs of all baselines and our model are available at this https URL.

Abstract (translated)

无监督文本样式转换的目的是传递文本的底层样式,但在没有并行数据的情况下保持其主要内容不变。大多数现有的方法通常遵循两个步骤:首先将内容与原始样式分离,然后将内容与所需样式融合。然而,第一步的分离是一个挑战,因为内容和风格在自然语言中以微妙的方式相互作用。因此,本文提出了一个双强化学习框架,通过一步映射模型直接传递文本的风格,而不分离内容和风格。具体来说,我们将源到目标和目标到源映射的学习视为一项双重任务,并基于这种双重结构设计了两个奖励,分别反映样式的准确性和内容的保存。这样,两个一步映射模型就可以通过强化学习来训练,而不需要使用任何并行数据。自动评估表明,我们的模型在很大程度上优于最先进的系统,特别是在两个基准数据集上平均改进了8个以上的bleu点。人的评价也验证了我们模型在风格准确性、内容保存和流畅性方面的有效性。我们的代码和数据,包括所有基线的输出和我们的模型都可以在这个https URL上找到。

URL

https://arxiv.org/abs/1905.10060

PDF

https://arxiv.org/pdf/1905.10060.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot