TeamEC at SemEval-2023 Task 4: Transformers vs. Low-Resource Dictionaries, Expert Dictionary vs. Learned Dictionary
Nicolas Stefanovitch, Bertrand De Longueville, Mario Scharfbillig
The 17th International Workshop on Semantic Evaluation (SemEval-2023) Task 4: valueeval: identification of human values behind arguments Paper
TLDR:
This paper describes the system we used to participate in the shared task, as well as additional experiments beyond the scope of the shared task, but using its data.Our primary goal is to compare the effectiveness of transformers model compared to low-resource dictionaries. Secondly, we compare the
You can open the
#paper-SemEval_317
channel in a separate window.
Abstract:
This paper describes the system we used to participate in the shared task, as well as additional experiments beyond the scope of the shared task, but using its data.Our primary goal is to compare the effectiveness of transformers model compared to low-resource dictionaries. Secondly, we compare the difference in performance of a learned dictionary and of a dictionary designed by experts in the field of values.Our findings surprisingly show that transformers perform on par with a dictionary containing less than 1k words, when evaluated with 19 fine-grained categories, and only outperform a dictionary-based approach in a coarse setting with 10 categories.Interestingly, the expert dictionary has a precision on par with the learned one, while its recall is clearly lower, potentially an indication of overfitting of topics to values in the shared task's dataset.Our findings should be of interest to both the NLP and Value scientific communities on the use of automated approaches for value classification