
This article explores the potential and limitations of Large Language Models (LLMs) in conducting qualitative thematic analysis (TA). By comparing GPT-4’s analysis with human-led analysis of a YouTube dataset on Roma migrants in Sweden, we examine both the computational efficiency and the interpretative constraints of AI-driven analysis. While GPT-4 demonstrates scalability in processing large datasets, its reliance on broad, neutral classifications overlooks the culturally embedded and ideologically charged dimensions of discourse. Drawing on perspectives that emphasize the situated nature of communication, we argue that human-AI synergy offers valuable methodological advancements but requires critical human oversight to ensure contextual depth and analytical precision. Our study underscores the need for refining LLM performance through culturally and politically informed training and theory-driven prompting to enhance their role in qualitative research.
ChatGPT-4, AI in qualitative research, human-AI synergy, Large Language Models (LLMs), thematic analysis (TA)
ChatGPT-4, AI in qualitative research, human-AI synergy, Large Language Models (LLMs), thematic analysis (TA)
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 2 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
