
From Pretraining Data to Language Models to Downstream Tasks: Tracking the Trails of Political Biases Leading to Unfair NLP Models
Even when pretraining data is clean, large language models can still absorb—and amplify—political bias. This study maps how that happens and how it affects hate speech and misinformation detection.