
We present a systematic empirical evaluation of sparse (TF-IDF), dense (MiniLM embeddings), and hybrid retrieval methods on a 19,965-document StackOverflow question–answer corpus. Retrieval performance is evaluated using Recall@K and Mean Reciprocal Rank (MRR) across 1,000 benchmark queries. Dense retrieval achieves Recall@5 = 0.779 and MRR = 0.670, significantly outperforming the sparse baseline (Recall@5 = 0.394, MRR = 0.292). A weighted hybrid method (α = 0.8) slightly improves Recall@5 to 0.789 and Recall@10 to 0.843 while marginally reducing ranking precision. These findings highlight trade-offs between recall coverage and ranking precision in domain-specific question answering systems.
Information Retrieval, Dense Retrieval, StackOverflow, Question Answering
Information Retrieval, Dense Retrieval, StackOverflow, Question Answering
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
