lnu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Unlocking language barriers: Assessing pre-trained large language models across multilingual tasks and unveiling the black box with Explainable Artificial Intelligence
University of New York Tirana, Albania.
Norwegian University of Science & Technology (NTNU), Norway.
Norwegian University of Science & Technology (NTNU), Norway.
Linnaeus University, Faculty of Technology, Department of Informatics.ORCID iD: 0000-0002-0199-2377
Show others and affiliations
2025 (English)In: Engineering applications of artificial intelligence, ISSN 0952-1976, E-ISSN 1873-6769, Vol. 149, article id 110136Article in journal (Refereed) Published
Abstract [en]

Large Language Models (LLMs) have revolutionized many industrial applications and paved the way for fostering a new research direction in many fields. Conventional Natural Language Processing (NLP) techniques, for instance, are no longer necessary for many text-based tasks, including polarity estimation, sentiment and emotion classification, and hate speech detection. However, training a language model for domain-specific tasks is hugely costly and requires high computational power, thereby restricting its true potential for standard tasks. This study, therefore, provides a comprehensive analysis of the latest pre-trained LLMs for various NLP-related applications without fine-tuning them to evaluate their effectiveness. Five language models are thus employed in this study on six distinct NLP tasks (including emotion recognition, sentiment analysis, hate speech detection, irony detection, offensiveness detection, and stance detection) for 12 languages from low- to medium- and high-resource. Generative Pre-trained Transformer 4 (GPT-4) and Gemini Pro outperform state-of-the-art models, achieving average F1 scores of 70.6% and 68.8% on the Tweet Sentiment Multilingual dataset compared to the state-of-the-art average F1 score of 66.8%. The study further interprets the findings obtained by the LLMs using Explainable Artificial Intelligence (XAI). To the best of our knowledge, it is the first time any study has employed explainability on pre-trained language models.

Place, publisher, year, edition, pages
Elsevier, 2025. Vol. 149, article id 110136
Keywords [en]
Large language models, Zero-shot classification, Explainable Artificial Intelligence, Sentiment analysis, Emotion recognition
National Category
Natural Language Processing
Research subject
Computer and Information Sciences Computer Science, Information Systems
Identifiers
URN: urn:nbn:se:lnu:diva-137213DOI: 10.1016/j.engappai.2025.110136ISI: 001446400200001Scopus ID: 2-s2.0-86000570396OAI: oai:DiVA.org:lnu-137213DiVA, id: diva2:1944090
Available from: 2025-03-12 Created: 2025-03-12 Last updated: 2025-03-27Bibliographically approved

Open Access in DiVA

fulltext(7519 kB)66 downloads
File information
File name FULLTEXT01.pdfFile size 7519 kBChecksum SHA-512
8a29a795d591cad4a84984bae01f9b03f40a700c0e98163483ac1e8afc39e3613a4a7826308a0530015f680d228f224c09a1abb96b33d7684abdab5d983bd2bd
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Kastrati, Zenun

Search in DiVA

By author/editor
Kastrati, Zenun
By organisation
Department of Informatics
In the same journal
Engineering applications of artificial intelligence
Natural Language Processing

Search outside of DiVA

GoogleGoogle Scholar
Total: 67 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 202 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf