Options
BERT Transformers Performance Comparison for Sentiment Analysis: A Case Study in Spanish
Journal
Good Practices and New Perspectives in Information Systems and Technologies : WorldCIST 2024
ISSN
2367-3370
Publisher
Springer
Date Issued
2024-01-01
Author(s)
Bárcena Ruiz, Gerardo
Gil Herrera, Richard de Jesús
Type
Resource Types::text::book::book part
Abstract
Despite the fact that the Spanish language is the second most spoken language in the world, research about AI and sentiment analysis is few compared with English language as research target. This paper refers to some works about sentiment analysis for the Spanish language that use BERT transformers and other technologies for this kind of sentiment analysis task; but the quality model based on indicators such as accuracy level could be different according to the tool or BERT version used. In addition, about the BERT family, it is challenging to determine which versions or subversions could perform better for sentiment analysis and also comply with the Spanish language when they are used on common platforms such as Colab. Therefore, the present study seeks to address this issue by establishing objectives, such as identifying relevant datasets based on the quality of Spanish used and having balanced subsets; also, locating different Spanish trained models; and proposing a method of comparison that involves relevant variables. We propose a weighted index that combines the F1-Score and the retraining time in different scenarios to help making better decisions. The results of this study indicate that the DistilBERT, RoBERTa, and ALBERT models have highest performances, but BERT remains in top positions as a consistent model. ©2024 Springer.
License
Acceso Restringido
How to cite
Bárcena Ruiz, G., de Jesús Gil, R. (2024). BERT Transformers Performance Comparison for Sentiment Analysis: A Case Study in Spanish. In: Rocha, Á., Adeli, H., Dzemyda, G., Moreira, F., Poniszewska-Marańda, A. (eds) Good Practices and New Perspectives in Information Systems and Technologies. WorldCIST 2024. Lecture Notes in Networks and Systems, vol 989. Springer, Cham. https://doi.org/10.1007/978-3-031-60227-6_13
