TY - GEN
T1 - CouRGe
T2 - 30th Irish Conference on Artificial Intelligence and Cognitive Science, AICS 2022
AU - Carraro, Diego
AU - Brown, Kenneth N.
N1 - Publisher Copyright:
© 2023, The Author(s).
PY - 2023
Y1 - 2023
N2 - Past literature in Natural Language Processing (NLP) has demonstrated that counterfactual data points are useful, for example, for increasing model generalisation, enhancing model interpretability, and as a data augmentation approach. However, obtaining counterfactual examples often requires human annotation effort, which is an expensive and highly skilled process. For these reasons, solutions that resort to transformer-based language models have been recently proposed to generate counterfactuals automatically, but such solutions show limitations. In this paper, we present CouRGe, a language model that, given a movie review (i.e. a seed review) and its sentiment label, generates a counterfactual review that is close (similar) to the seed review but of the opposite sentiment. CouRGe is trained by supervised fine-tuning of GPT-2 on a task-specific dataset of paired movie reviews, and its generation is prompt-based. The model does not require any modification to the network’s architecture or the design of a specific new task for fine-tuning. Experiments show that CouRGe’s generation is effective at flipping the seed sentiment and produces counterfactuals reasonably close to the seed review. This proves once again the great flexibility of language models towards downstream tasks as hard as counterfactual reasoning and opens up the use of CouRGe’s generated counterfactuals for the applications mentioned above.
AB - Past literature in Natural Language Processing (NLP) has demonstrated that counterfactual data points are useful, for example, for increasing model generalisation, enhancing model interpretability, and as a data augmentation approach. However, obtaining counterfactual examples often requires human annotation effort, which is an expensive and highly skilled process. For these reasons, solutions that resort to transformer-based language models have been recently proposed to generate counterfactuals automatically, but such solutions show limitations. In this paper, we present CouRGe, a language model that, given a movie review (i.e. a seed review) and its sentiment label, generates a counterfactual review that is close (similar) to the seed review but of the opposite sentiment. CouRGe is trained by supervised fine-tuning of GPT-2 on a task-specific dataset of paired movie reviews, and its generation is prompt-based. The model does not require any modification to the network’s architecture or the design of a specific new task for fine-tuning. Experiments show that CouRGe’s generation is effective at flipping the seed sentiment and produces counterfactuals reasonably close to the seed review. This proves once again the great flexibility of language models towards downstream tasks as hard as counterfactual reasoning and opens up the use of CouRGe’s generated counterfactuals for the applications mentioned above.
KW - Counterfactual reasoning
KW - Data augmentation
KW - Language models
KW - Natural language processing
KW - Sentiment analysis
UR - https://www.scopus.com/pages/publications/85149987733
U2 - 10.1007/978-3-031-26438-2_24
DO - 10.1007/978-3-031-26438-2_24
M3 - Conference proceeding
AN - SCOPUS:85149987733
SN - 9783031264375
T3 - Communications in Computer and Information Science
SP - 305
EP - 317
BT - Artificial Intelligence and Cognitive Science - 30th Irish Conference, AICS 2022, Revised Selected Papers
A2 - Longo, Luca
A2 - O’Reilly, Ruairi
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 8 December 2022 through 9 December 2022
ER -