Can AI Detect AI in GCSE Coursework Reliably?
The question every GCSE teacher is quietly asking is whether AI detection is reliable enough to actually do anything with. A student submits a suspiciously polished controlled assessment. You run it through a detection tool. A likelihood score appears. And then what? Can you act on it? Can you challenge the student? Can you flag it to a senior leader?
The honest answer is: it depends on what you mean by "reliable." AI detection tools — including the one built into GradeOrbit — are genuinely useful, but only when teachers understand what they are actually measuring. A likelihood score is probabilistic data, not a verdict. This guide explains what that means in practice, where the limits lie, and how teachers can use detection output from tools like ChatGPT and Claude to support their professional judgment rather than replace it.
What a Likelihood Score Actually Tells You
When a detection tool analyses a piece of student writing, it compares the linguistic patterns in that text against a statistical model built from large samples of both human and AI-generated writing. The output — expressed as a percentage from 0 to 100% — represents the tool's confidence that the text was produced by an AI model such as ChatGPT or Claude.
A score of 80% does not mean "80% of this essay was written by AI." It means the model considers the text to be highly consistent with AI-generated writing based on the patterns it has learned. A score of 20% means the writing looks largely human. Scores in the 40–65% range are genuinely ambiguous and should be treated with particular caution.
What the score cannot tell you is who wrote what, when, or how much editing took place after the fact. A student who generated a first draft using ChatGPT and then rewrote large sections in their own voice might score anywhere from 30% to 90% depending on how thoroughly they edited. A student with unusually sophisticated writing for their age — or one who writes in a very structured, formal style — may score higher than you would expect despite having produced every word themselves.
This is not a flaw in the tool. It is the honest nature of probabilistic analysis. The score is a signal, not a conclusion.
Why the Same Work Can Score Differently on 1 vs 3 Credits
GradeOrbit offers two detection models: a quick scan using 1 credit and a deep scan using 3 credits. The quick scan provides a rapid initial assessment — it processes the text against a core detection model and returns a likelihood score in seconds. This is useful for initial screening when you have a batch of coursework and want to identify pieces that warrant closer attention.
The deep scan runs a more thorough multi-pass analysis. It looks at longer segments of text, analyses internal consistency across the document, and applies additional detection layers that the quick scan does not use. For this reason, the deep scan is generally more reliable on longer pieces of work, and you may see a different score from the deep scan than from the quick scan on the same document.
Neither result should be read as a definitive answer in isolation. The right approach is to use the quick scan as a filter and the deep scan for pieces where you have already formed a view that something may be wrong. Treat both scores as one input among several rather than the final word on authenticity.
The Types of Writing AI Detection Struggles With
Detection tools are not equally reliable across all genres of student writing. There are specific categories of work where scores are less meaningful and should be weighted accordingly.
EAL Students
Students who are learning English as an additional language often produce writing that is syntactically simpler, less idiomatic, and more formulaic than their fluent peers. Detection models have largely been trained on native-English writing and may misread the patterns in EAL work. A student whose grammar follows more rigid, textbook-influenced structures may produce a higher likelihood score simply because their writing does not match the irregular, natural patterns of a native English speaker. This is a known limitation and one that teachers should weigh carefully before drawing any conclusions from a detection score on work submitted by an EAL student.
Structured and Formulaic Genres
Certain types of GCSE writing are inherently formulaic. Science coursework write-ups, GCSE Geography fieldwork reports, and Business Studies case study analyses all follow predictable templates that students are explicitly taught to replicate. Because AI models were also trained on examples of these genres, the patterns overlap significantly with human-written work in the same format. A student who has learned the structure well and applied it correctly might produce work that reads similarly to AI output — not because they used AI, but because they mastered the genre they were taught.
Short Answers and Single Paragraphs
Detection tools work best on extended writing with sufficient volume for statistical patterns to emerge. Running detection on a single paragraph or a short-answer response of fewer than 150 words is unlikely to produce a meaningful score. The sample size is too small for the model to identify reliable patterns. For shorter pieces of work, qualitative teacher judgment is more valuable than detection scores.
How to Use Scores as a Conversation Starter, Not a Verdict
The most important principle in using AI detection professionally is that a score never justifies a direct accusation on its own. Even a score of 95% should be treated as a reason to look more closely, not as proof that misconduct occurred.
The practical approach that most experienced teachers find effective is to use detection scores to prioritise which pieces of work deserve more careful scrutiny — and then apply human judgment to those pieces. That scrutiny might involve comparing the flagged submission to the student's classwork or previous assessments, checking whether the writing style is consistent with how the student performs in supervised conditions, or having a brief, low-pressure conversation where you ask the student to talk through their work.
A student who wrote their coursework genuinely can almost always elaborate on it verbally. They can describe decisions they made, explain why they chose one approach over another, or articulate something they found difficult. Students who outsourced the writing to ChatGPT or Claude typically struggle to go beyond what is on the page, because the knowledge behind the words was never theirs.
This approach — score as signal, conversation as investigation — is not only fairer to students, it is also more robust if a concern ever needs to be escalated. A detection score plus a record of a student who could not explain their own work is a far stronger foundation for a formal discussion than a score alone.
For more detail on navigating difficult conversations after detection, see the guide on talking to students about AI detection results.
GradeOrbit's Built-In Detection — What You Get
GradeOrbit's AI detection tool is designed specifically for UK teachers, not as a standalone product but as part of the same workflow you already use for marking. You can run detection on the same uploaded work you are about to mark, or submit work specifically for detection without going through the full marking process.
The tool produces a clear likelihood score from 0 to 100% alongside a supporting analysis that highlights the sections of text that contributed most to the score. This gives you something concrete to look at rather than a single number with no context. You can see which paragraphs flagged highest, which sections appeared most consistent with AI-generated patterns, and where the writing looked most genuinely human.
Crucially, nothing is stored. Student work processed through GradeOrbit is never saved to a database or used to train any model. Every upload is processed in the moment and then discarded. This makes GradeOrbit safe to use under UK GDPR without needing to seek special permissions or notify parents — the data simply does not persist.
The credit system is straightforward: 1 credit for a quick scan, 3 credits for a deep scan. There is no subscription requirement and no minimum commitment. You use GradeOrbit when you need it.
Try AI Detection Built for Teachers
AI detection is a genuinely useful tool in the professional toolkit of a UK secondary school teacher — but only when it is understood for what it is. It provides probabilistic signals that support professional judgment, not automated verdicts that replace it. Used thoughtfully, a good detection tool helps you focus your attention, ask better questions, and make fairer, better-evidenced decisions about student work.
GradeOrbit's built-in detection is designed with exactly this purpose in mind. It is fast, private, and built around the realities of the UK classroom — including handwritten work, formulaic genres, and the professional responsibilities that come with any assessment decision.
Create a free GradeOrbit account and start using AI detection the way it was meant to be used: as a tool that informs your judgment, not one that makes it for you.