CardiffNLP-Metaphor at SemEval-2022 Task 2: Targeted Fine-tuning of Transformer-based Language Models for Idiomaticity Detection

Abstract

This paper describes the experiments ran for SemEval-2022 Task 2, subtask A, zero-shot and one-shot settings for idiomaticity detection. Our main approach is based on fine-tuning transformer-based language models as a baseline to perform binary classification. Our system, CardiffNLP-Metaphor, ranked 8th and 7th (respectively on zero- and one-shot settings on this task. Our main contribution lies in the extensive evaluation of transformer-based language models and various configurations, showing, among others, the potential of large multilingual models over base monolingual models. Moreover, we analyse the impact of various input parameters, which offer interesting insights on how language models work in practice.

Type
Publication
Proceedings of the 16th International Workshop on Semantic Evaluation (SemEval-2022)
Joanne Boisson
Joanne Boisson
PhD Student & Senior Machine Learning Engineer Amplyfi
Jose Camacho-Collados
Jose Camacho-Collados
Professor & UKRI Future Leaders Fellow
Luis Espinosa-Anke
Luis Espinosa-Anke
Senior Lecturer