From Pretraining Data to Language Models to Downstream Tasks: Tracking the Trails of Political Biases Leading to Unfair NLP Models https://aclanthology.org/2023.acl-long.656.pdf
From Pretraining Data to Language Models to Downstream Tasks: Tracking the Trails of Political Biases Leading to Unfair NLP Models https://aclanthology.org/2023.acl-long.656.pdf
> The models that were trained with left-wing data were more sensitive to hate speech targeting ethnic, religious, and sexual minorities in the US, such as Black and LGBTQ+ people. The models that were trained on right-wing data were more sensitive to hate speech against white Christian men.
White christian men is an awfully specific thing for the model to be sensitive towards IMO.
Right-wing media is perceived to be funded by white christian men, so if that is the source of the data then I’m not too surprised their writing and articles would protect themselves - but still intriguing how the model picked up on this from online discussions & news data, and was sensitive to hate speech aimed at that group specifically, compared with the Left data which appears more inclusive - although this is probably indicative of the bias they’re studying in the article
I mean, hate speech aimed at left-wing people is more diverse generally than hate speech aimed at right-wing people because the left simply is more diverse in gender, orientation, ethnicity, religion, etc. Isn’t that universally accepted?
(Please correct me if I’m wrong, I approach in good faith!)
I don’t think you’re wrong at all tbh - from my perspective the left is always going to be more diverse, whereas the right isn’t very inclusive by default unless you “fit in” IMO