@inproceedings{shuo-2022-tagging,
title = "Tagging Without Rewriting: A Probabilistic Model for Unpaired Sentiment and Style Transfer",
author = "Yang, Shuo",
editor = "Barnes, Jeremy and
De Clercq, Orph{\'e}e and
Barriere, Valentin and
Tafreshi, Shabnam and
Alqahtani, Sawsan and
Sedoc, Jo{\~a}o and
Klinger, Roman and
Balahur, Alexandra",
booktitle = "Proceedings of the 12th Workshop on Computational Approaches to Subjectivity, Sentiment {\&} Social Media Analysis",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2022.wassa-1.33/",
doi = "10.18653/v1/2022.wassa-1.33",
pages = "293--303",
abstract = "Style transfer is the task of paraphrasing text into a target-style domain while retaining the content. Unsupervised approaches mainly focus on training a generator to rewrite input sentences. In this work, we assume that text styles are determined by only a small proportion of words; therefore, rewriting sentences via generative models may be unnecessary. As an alternative, we consider style transfer as a sequence tagging task. Specifically, we use edit operations (i.e., deletion, insertion and substitution) to tag words in an input sentence. We train a classifier and a language model to score tagged sequences and build a conditional random field. Finally, the optimal path in the conditional random field is used as the output. The results of experiments comparing models indicate that our proposed model exceeds end-to-end baselines in terms of accuracy on both sentiment and style transfer tasks with comparable or better content preservation."
}
Markdown (Informal)
[Tagging Without Rewriting: A Probabilistic Model for Unpaired Sentiment and Style Transfer](https://preview.aclanthology.org/jlcl-multiple-ingestion/2022.wassa-1.33/) (Yang, WASSA 2022)
ACL