Definition
The "AI as Expert Witness Fallacy" refers to the mistaken belief that large language models can weigh evidence or reason like humans, leading to their inappropriate use as authoritative sources for reasoned conclusions. According to Steve Hargadon, this represents a fundamental misunderstanding of how large language models function and constitutes both a technical and ethical problem in the application of AI technology.
Mechanism and Core Problem
Hargadon explains that large language models calculate responses based on the frequency of language patterns, where "the prevalence of opinions—especially on contentious topics—often has little to do with actual truth." The fundamental issue is that these systems cannot evaluate information for truthfulness, despite their ability to assert that something is true. Instead, they are "merely mimicking human claims of truth."
The fallacy emerges when users feed LLMs articles that support a particular position and ask the system to craft a response based on them. The model will reflect that input, "essentially echoing the narrative you've curated." This selective feeding creates what Hargadon describes as "a kind of echo chamber, where the output feels authoritative but is just a snapshot of the provided data, not a broader truth."
The "Stochastic Parrots" Concept
Drawing on the concept of "stochastic parrots," Hargadon characterizes large language models as systems that "predict and string together words based on statistical patterns, not understanding or critical thinking." This technical limitation means that while LLMs can produce responses that appear reasoned, they lack the capacity for genuine evaluation or critical judgment that would qualify them as expert witnesses.
The "Reasoning Models" Deception
Hargadon addresses the marketing of certain AI systems as "reasoning models," arguing that these represent advanced pattern-matching rather than genuine reasoning. He contends that even these sophisticated models are "doing an impressive job of identifying patterned questions and recalculating responses based on new guidelines. It looks like reasoning, but it's not what we consider human reasoning—no extrapolation or critical judgment is happening."
According to Hargadon, providers calling these "reasoning models" can mislead users into thinking they're receiving independent insight when the output represents "just advanced pattern-matching."
Ethical Implications
Hargadon frames the misuse of AI as expert witness as fundamentally ethical, not merely technical. He identifies several concerns: AI can amplify biases from its training data, and it can be used to manipulate or deceive when treated as a trusted source. This misapplication "underscores the need for caution" in how AI systems are positioned and utilized.
Educational and Cognitive Risks
The fallacy poses particular risks in educational contexts. Hargadon notes concerns about "widespread student use of AI and its apparent reasoning," suggesting this "could signal a growing problem for critical thinking." He warns that treating AI like an expert witness or historian "risks undermining our ability to question and reason for ourselves."
Hargadon draws a parallel to over-reliance on Wikipedia, suggesting that using AI as a final authority rather than a starting point represents a similar abdication of critical thinking responsibility.
Appropriate AI Applications
Despite these limitations, Hargadon emphasizes his appreciation for large language models in their proper applications. He identifies specific areas where LLMs excel: research and surfacing information quickly, such as "synthesizing trends in discussions about digital literacy or pulling together studies for a literature review." He also highlights their value for "research, brainstorming, and spotting patterns."
Recommended Framework
Hargadon's solution involves a clear division of labor: "We need to use AI for what it's great at—research, brainstorming, and spotting patterns—while reserving judgment and truth-seeking for human minds." This framework acknowledges the substantial utility of AI systems while maintaining appropriate boundaries around their authoritative use.
The core principle underlying this approach is recognition that while AI can efficiently process and synthesize information, the evaluation of that information's truthfulness and the drawing of reasoned conclusions remain distinctly human capabilities that should not be delegated to pattern-matching systems, regardless of their sophistication.