With the development of social networks, there has been a signicant in-
crease of hate speech in user generated contents. We focus on two most
discussed topics, LGBT and migrants. We use the BERT neural network for
prediction of hate speech and make a comparison between the multilingual
model, trained on 104 dierent languages, and a trilingual model, trained on
Slovene, Croatian and English. Results show that the trilingual model is ap-
proximately 5% more accurate predicting hate speech on a language that it
was trained on. The multilingual model with or without additional training
is more accurate on languages that it was not trained on. This indicates a
better cross-lingual transfer of multilingual model.
|