Details

Odkrivanje pristranosti in razpristranjevanje v jezikovnem modelu SloBERTa
ID Narat, Marko (Author), ID Robnik Šikonja, Marko (Mentor) More about this mentor... This link opens in a new window

.pdfPDF - Presentation file, Download (611,59 KB)
MD5: 0C7C23B9FEDD262E5732BA9676D4469D

Abstract
Veliki jezikovni modeli so omogočili velik napredek pri obdelavi naravnega jezika, vendar so lahko pristrani, predvsem zaradi učnih podatkov. Pristranost lahko vodi do diskriminatornih odločitev v zaposlovanju, v pravosodju in v drugih aplikacijah. V magistrskem delu raziskujemo spolno pristranost slovenskega maskirnega jezikovnega modela SloBERTa. Predstavimo in uporabimo šest metod za analizo pristranosti ter prilagodimo angleško metodo razpristranjevanja Kaneko in Bollegala (2021) za slovenski model SloBERTa. Zmogljivost modela pred in po razpristranjevanju ovrednotimo z nalogami iz zbirke SuperGLUE ter z nalogo klasifikacije sentimenta z množico Sentinews. Rezultati testa stavčne asociacije potrjujejo pristranost v prid moškim v odnosu do poklicev v modelu SloBERTa, medtem ko ostalih pet metod pristranosti test ne zazna. Po razpristranjevanju pristranosti ne zazna več nobena metoda, vendar razpristranjeni modeli dosegajo slabše rezultate pri nalogah iz zbirke SuperGLUE in pri nalogi klasifikacije sentimenta.

Language:Slovenian
Keywords:veliki jezikovni modeli, pristranost, razpristranjevanje, SloBERTa, spolna pristranost
Work type:Master's thesis/paper
Typology:2.09 - Master's Thesis
Organization:PEF - Faculty of Education
Publication status:Published
Publication version:Version of Record
Place of publishing:Ljubljana
Publisher:M. Narat
Year:2025
Number of pages:68 str.
PID:20.500.12556/RUL-167343 This link opens in a new window
UDC:81(043.2)
COBISS.SI-ID:226284547 This link opens in a new window
Publication date in RUL:16.02.2025
Views:502
Downloads:100
Metadata:XML DC-XML DC-RDF
:
Copy citation
Share:Bookmark and Share

Secondary language

Language:English
Title:Detecting and removing biases in the SloBERT model
Abstract:
Large language models have significantly advanced the area of natural language processing, but they inherit biases present in the training data. The learned biases can lead to discriminatory decisions in employment, justice, and other applications. This work explores gender bias in the Slovene masked large language model SloBERTa. We present and apply six methods for bias analysis and adapt the English debiasing method Kaneko and Bollegala (2021) for the Slovenian SloBERTa model. The model's performance before and after debiasing is evaluated using tasks from the SuperGLUE benchmark and sentiment classification task with the Sentinews dataset. The Sentence Association Test results confirm a male bias in relation to professions in the original SloBERTa model, while the other five bias detection methods do not identify such bias. After debiasing, none of the methods detect bias. However, the debiased models show a decline in performance on selected tasks from the SuperGLUE benchmark and the sentiment classification task.

Keywords:Large language models, bias, debiasing, SloBERTa, gender bias

Similar documents

Similar works from RUL:
Similar works from other Slovenian collections:

Back