If AI Can’t Guarantee Truth, Why Trust It to Spot Plagiarism?

Headshot
December 2, 2024

If we know that tools like ChatGPT generate outputs that are often inaccurate, incomplete, or fabricated, why are we so quick to believe that AI detection tools—built on the same underlying technology—are any more reliable? This question sits at the heart of a growing dilemma in academia and beyond, where generative AI and AI-driven detection systems are being adopted without sufficient scrutiny.

The problem begins with a misplaced trust in what these systems do. Generative AI models like ChatGPT don’t “know” anything; they generate text based on probabilities derived from their training data. Their outputs are coherent but often wrong, a phenomenon known as hallucination. Yet, when it comes to detection tools—designed to flag AI-generated content or determine authorship—this basic truth seems to be forgotten. Institutions treat these systems as reliable arbiters of originality, even though they are governed by the same probabilistic mechanisms as the generative models themselves.

The Flaws in AI Detection

AI detectors work by analysing patterns in text—sentence structures, word choices, and stylistic markers—that align with known characteristics of machine-generated content. But like generative AI, detection tools rely on statistical probabilities, not definitive truths. This makes them prone to false positives, misclassifying human-generated text as AI-written.

Non-native English speakers, for instance, often write in ways that deviate from standard conventions embedded in training data, making their work more likely to be flagged. Similarly, creative or unconventional writing styles can confuse detection systems, resulting in false accusations of plagiarism or misconduct. These errors carry real consequences—damaged reputations, academic penalties, and a loss of trust between institutions and individuals.

The irony is hard to ignore: we are deploying AI systems to identify AI-generated content, all while acknowledging that these same systems are unreliable in producing factual or accurate outputs.

Why Are We So Quick to Trust AI Detectors?

Part of the problem lies in perception. AI detection tools are marketed as solutions to the challenges posed by generative AI, offering an illusion of control in a rapidly shifting landscape. They promise certainty where ambiguity reigns. But this trust is misplaced.

The reality is that AI detectors share the same limitations as generative models. They are trained on incomplete, biased, and often ethically questionable datasets. They cannot contextualise or evaluate intent. And like all statistical systems, they are only as reliable as the data and algorithms behind them—data and algorithms that remain opaque to most users.

This misplaced trust reflects a broader societal trend: the willingness to defer to technology as an objective authority, even when the stakes involve human judgment. In academic settings, this deference can erode the values of fairness and integrity that institutions are meant to uphold.

The Broader Implications

The implications extend far beyond false plagiarism accusations. When institutions rely on AI detection tools, they risk creating a chilling effect on creativity and diversity in writing. Students, aware of the risks of misclassification, may feel pressured to conform to rigid linguistic norms to avoid suspicion. This not only stifles originality but also penalises those whose voices don’t align with algorithmic expectations.

Moreover, reliance on AI detection tools shifts the focus from teaching and learning to policing. Instead of fostering an environment where students develop critical thinking and writing skills, institutions are incentivised to adopt a defensive posture, prioritising detection over education.

Moving Beyond the Illusion of Certainty

If we can’t trust generative AI to provide accurate outputs, we shouldn’t trust detection tools to reliably identify them. Both systems operate within the same probabilistic framework, and neither can deliver the level of certainty that their users often expect.

The way forward requires a fundamental shift in how we approach these technologies. Institutions need to recognise the limitations of AI and resist the temptation to delegate human judgment to machines. Educators must remain central to decisions about originality, intent, and integrity, using AI as a tool to inform—not dictate—their assessments.

Transparency is equally crucial. Institutions must educate students, staff, and administrators about the limitations of AI detection tools, including their susceptibility to error and bias. Without this openness, trust in these systems will continue to erode, leaving behind a landscape of suspicion and inequity.

At its core, this issue isn’t about technology; it’s about values. Are we willing to accept the inherent uncertainties of human creativity, or will we allow flawed systems to dictate what counts as original or authentic? The answer will shape not only how we use AI but also how we define fairness and integrity in an increasingly automated world.