ZeroBERTo: Leveraging Zero-Shot Text Classification by Topic Modeling - Télécom Paris Accéder directement au contenu
Chapitre D'ouvrage Année : 2022

ZeroBERTo: Leveraging Zero-Shot Text Classification by Topic Modeling

Résumé

Traditional text classification approaches often require a good amount of labeled data, which is difficult to obtain, especially in restricted domains or less widespread languages. This lack of labeled data has led to the rise of low-resource methods, that assume low data availability in natural language processing. Among them, zero-shot learning stands out, which consists of learning a classifier without any previously labeled data. The best results reported with this approach use language models such as Transformers, but fall into two problems: high execution time and inability to handle long texts as input. This paper proposes a new model, ZeroBERTo, which leverages an unsupervised clustering step to obtain a compressed data representation before the classification task. We show that ZeroBERTo has better performance for long inputs and shorter execution time, outperforming XLM-R by about 12% in the F1 score in the FolhaUOL dataset.
Fichier principal
Vignette du fichier
preprint_zeroberto.pdf (447.61 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03628242 , version 1 (04-06-2022)

Identifiants

Citer

Alexandre Alcoforado, Thomas Palmeira Ferraz, Rodrigo Gerber, Enzo Bustos, André Seidel Oliveira, et al.. ZeroBERTo: Leveraging Zero-Shot Text Classification by Topic Modeling. Vládia Pinheiro; Pablo Gamallo; Raquel Amaro; Carolina Scarton; Fernando Batista; Diego Silva; Catarina Magro; Hugo Pinto. Computational Processing of the Portuguese Language. 15th International Conference, PROPOR 2022, Fortaleza, Brazil, March 21–23, 2022, Proceedings, 13208, Springer International Publishing, pp.125-136, 2022, Lecture Notes in Computer Science, 978-3-030-98304-8. ⟨10.1007/978-3-030-98305-5_12⟩. ⟨hal-03628242⟩
37 Consultations
258 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More