Title Low Resource Style Transfer via Domain Adaptive Meta Learning
Authors Li, Xiangyang
Long, Xiang
Xia, Yu
Li, Sujian
Affiliation Peking Univ, Key Lab Computat Linguist, MOE, Beijing, Peoples R China
Beijing Univ Posts & Telecommun, Beijing, Peoples R China
Issue Date 2022
Publisher NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES
Abstract Text style transfer (TST) without parallel data has achieved some practical success. However, most of the existing unsupervised text style transfer methods suffer from (i) requiring massive amounts of non-parallel data to guide transferring different text styles. (ii) colossal performance degradation when fine-tuning the model in new domains. In this work, we propose DAML-ATM (Domain Adaptive Meta-Learning with Adversarial Transfer Model), which consists of two parts: DAML and ATM. DAML is a domain adaptive meta-learning approach to learn general knowledge in multiple heterogeneous source domains, capable of adapting to new unseen domains with a small amount of data. Moreover, we propose a new unsupervised TST approach Adversarial Transfer Model (ATM), composed of a sequence-to-sequence pre-trained language model and uses adversarial style training for better content preservation and style transfer. Results on multi-domain datasets demonstrate that our approach generalizes well on unseen low-resource domains, achieving state-of-the-art results against ten strong baselines.
URI http://hdl.handle.net/20.500.11897/657179
ISBN 978-1-955917-71-1
DOI 10.48550/arXiv.2205.12475
Indexed EI
CPCI-SSH(ISSHP)
CPCI-S(ISTP)
Appears in Collections: 计算语言学教育部重点实验室

Files in This Work
There are no files associated with this item.

Web of Science®


0

Checked on Last Week

Scopus®



Checked on Current Time

百度学术™


0

Checked on Current Time

Google Scholar™





License: See PKU IR operational policies.