
Abstract This paper investigates the presence of political bias in emotion inference models used for sentiment analysis (SA). Machine learning models often reflect biases in their training data, impacting the validity of their outcomes. While previous research has highlighted gender and race biases, our study focuses on political bias—an underexplored, pervasive issue that can skew the interpretation of text data across many studies. We audit a Polish sentiment analysis model developed in our lab for bias. By analyzing valence predictions for names and sentences involving Polish politicians, we uncovered systematic differences influenced by political affiliations. Our findings suggest that annotations by human raters propagate political biases into the model’s predictions. To prove it, we pruned the training dataset of texts mentioning these politicians and observed a reduction in bias, though not its complete elimination. Given the significant implications of political bias in SA, our study emphasizes caution in employing these models for social science research. We recommend a critical examination of SA results and propose using lexicon-based systems as an ideologically neutral alternative. This paper underscores the necessity for ongoing scrutiny and methodological adjustments to ensure the reliability of the use of machine learning in academic and applied contexts.
FOS: Computer and information sciences, Social Science Research, Political Bias, Computer Science - Computation and Language, Computer Science - Artificial Intelligence, Science, Q, Politics, Emotions, R, Emotion inference models, Models, Theoretical, Article, Annotator Bias, Machine Learning, Sentiment analysis, Artificial Intelligence (cs.AI), Bias, Medicine, Humans, Computation and Language (cs.CL)
FOS: Computer and information sciences, Social Science Research, Political Bias, Computer Science - Computation and Language, Computer Science - Artificial Intelligence, Science, Q, Politics, Emotions, R, Emotion inference models, Models, Theoretical, Article, Annotator Bias, Machine Learning, Sentiment analysis, Artificial Intelligence (cs.AI), Bias, Medicine, Humans, Computation and Language (cs.CL)
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
