Evaluating Bias Detection in Lightweight LLMs
Straipsniai
Veronika Bryskina
Vytauto Didžiojo universitetas image/svg+xml
Milita Songailaitė
Vytauto Didžiojo universitetas image/svg+xml
Justina Mandravickaitė
Vytauto Didžiojo universitetas image/svg+xml
Publikuota 2026-05-08
https://doi.org/10.15388/LMITT.2026.4
PDF

Anotacija

This study evaluates the ability of lightweight open-source large language models (LLMs) to detect bias in text. Eleven models of six popular LLM families were tested in a zero-shot setting on a unified dataset of 8,745 sentences derived from three selected sources, covering gender, race, religion, and appearance bias. Results showed that none of the models exceeded 70% accuracy, which highlighted limitations of lightweight LLMs and existing challenges related to current bias detection datasets.

PDF

Nuorodos

Creative Commons License

Šis darbas apsaugotas Creative Commons priskyrimo 4.0 viešąja licencija.

Atsisiuntimai

Nėra atsisiuntimų.

Dažniausiai skaitomi to paties autoriaus (-ių) straipsniai