This article introduces DelBERTo, a resource-efficient Transformer architecture for Natural Language Processing (NLP). Transformers replace convolutions and recurrence with the self-attention mechanism and represent the state-of-the-art in NLP. However, self-attention’s complexity grows quadratically with the size of the input, which limits their applications. DelBERTo relies on adaptive input and on a deep yet lightweight Transformer architecture to reduce the number of learnable parameters, and relies on adaptive softmax to improve pre-training speed and memory footprint. We evaluate the proposed architecture in a sentiment analysis task and compare it against AlBERTo, a BERT model representing the state-of-the-art in sentiment analysis over Italian tweets. DelBERTo has only one-seventh of AlBERTo’s learnable parameters, is faster, and requires less memory. Despite this, our experiments show that DelBERTo is competitive with AlBERTo over the three SENTIPOLC sub-tasks proposed at EVALITA 2016: subjectivity classification, polarity classification, and irony detection.

DelBERTo: A Deep Lightweight Transformer for Sentiment Analysis

Luca Molinaro
First
;
Attilio Fiandrotti
;
Valerio Basile;Viviana Patti
2023-01-01

Abstract

This article introduces DelBERTo, a resource-efficient Transformer architecture for Natural Language Processing (NLP). Transformers replace convolutions and recurrence with the self-attention mechanism and represent the state-of-the-art in NLP. However, self-attention’s complexity grows quadratically with the size of the input, which limits their applications. DelBERTo relies on adaptive input and on a deep yet lightweight Transformer architecture to reduce the number of learnable parameters, and relies on adaptive softmax to improve pre-training speed and memory footprint. We evaluate the proposed architecture in a sentiment analysis task and compare it against AlBERTo, a BERT model representing the state-of-the-art in sentiment analysis over Italian tweets. DelBERTo has only one-seventh of AlBERTo’s learnable parameters, is faster, and requires less memory. Despite this, our experiments show that DelBERTo is competitive with AlBERTo over the three SENTIPOLC sub-tasks proposed at EVALITA 2016: subjectivity classification, polarity classification, and irony detection.
2023
22nd confernce of the Associazione Italiana per l'Intelligenza Artificiale - AIXIA
Udine
November 28 - December 2, 2022
Proceedings of the 22nd confernce of the Associazione Italiana per l'Intelligenza Artificiale - AIXIA
springer
443
456
978-3-031-27180-9
https://link.springer.com/chapter/10.1007/978-3-031-27181-6_31
Efficient transformer, Sustainable NLP, Sentiment analysis
Luca Molinaro, Rosalia Tatano, Enrico Busto, Attilio Fiandrotti, Valerio Basile, Viviana Patti
File in questo prodotto:
File Dimensione Formato  
978-3-031-27181-6_31.pdf

Accesso aperto

Tipo di file: PDF EDITORIALE
Dimensione 554.95 kB
Formato Adobe PDF
554.95 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2318/1945362
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 1
social impact