Fine-tuning multilingual transformers for Hinglish sentiment analysis: a comparative evaluation with BiLSTM

International Journal of Artificial Intelligence

Fine-tuning multilingual transformers for Hinglish sentiment  analysis: a comparative evaluation with BiLSTM

Abstract

Growing trend of code-mixing in languages, in the form of Hinglish, greatly tests the skills of conventional sentiment analysis tools. The research contributes a fine-tuned multilingual transformer model built exclusively for classifying sentiment of Hinglish customer reviews. Drawing from pre trained BERT-base-multilingual-case architecture, the model gets transformed with the process of fine-tuning the same on synthetically prepared and balanced dataset simulating positive, negative, and neutral sentiments. Sophisticated methods like focal loss for addressing the class imbalance and mixed precision training for maximization of computational effectiveness are embedded within the training process. Experimental results suggest that the proposed method significantly captures the fine-grained linguistic patterns of code-mixed text, improving sentiment classification accuracy. The results show promising potential for enhancing customer feedback analysis in e-commerce, social media monitoring, and customer support use cases, where it is crucial to comprehend the sentiment behind code-mixed reviews.

Discover Our Library

Embark on a journey through our expansive collection of articles and let curiosity lead your path to innovation.

Explore Now
Library 3D Ilustration