
This keynote addresses two key questions: Are large language models (LLMs) the right interface for data and information access, and what harms do AI models pose to institutions like libraries today? He explains that current LLMs, while powerful, are fundamentally flawed pattern-matchers rather than true reasoning systems. New techniques like chain-of-thought processing, self-verification, and retrieval-augmented generation (RAG) offer partial improvements but rely on the same unreliable foundations. On the second question, Zhao highlights the growing problem of AI-driven web scraping, noting that most mitigation strategies offer limited protection. He concludes that today’s generative AI LLMs are fundamentally flawed, and while composition techniques offer limited improvement, meaningful progress will require new architectures built with better understanding and ethical data sourcing. In the meantime, AI-driven crawlers pose an immediate threat, with most conventional defences proving ineffective—leaving commercial, network-level blocking as one of the few viable mitigation strategies.View on YouTube: https://www.youtube.com/watch?v=HL62C3U8epE
AI Bots, Access to Information, Large Language Models, Generative AI, OR2025
AI Bots, Access to Information, Large Language Models, Generative AI, OR2025
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
