Most AI checking tools claim 95-99%, but independent tests put the real rate at 70-94% depending on the tool and content type. False positive rates - where human writing gets wrongly flagged - range from under 4% (Turnitin) to over 26% (ZeroGPT). AI Busted tests five major tools so you know which ones to trust and which to skip.
What Are AI Detectors?
AI checking tools scan text for patterns that suggest machine authorship. The two main signals they measure are perplexity - how repetitive the word choices are across a passage - and burstiness - how much sentence length varies. AI text tends to score low on both. The fundamental limitation is structural. These tools were built using AI output from specific models at specific points in time. As AI writing improves and as humanizing tools spread, what was once a useful signal becomes noise. That's why vendor numbers from 2023 look very different from independent results run in 2026. For a plain-language breakdown of how the process works, see what is AI content checking.What Does Each Tool Claim vs. What Do Tests Show?
No vendor publishes this comparison. Here it is: vendor claim, real-world catch rate, and false positive rate side by side.| Tool | Vendor Claim | Independent Rate | False Positive Rate | Best For |
|---|---|---|---|---|
| Turnitin | 98% | ~94% | 3.8-6% (ESL: 6-8%) | Institutions, long-form academic papers |
| Originality.ai | 99%+ | 79-92% | 0.5-5.7% | Publishers, content teams |
| Copyleaks | <0.2% FP | 77% | 5-12% | General use; caution with ESL writers |
| GPTZero | 99.3% | 76-88% | 1-18% (ESL: 61%) | Education; unreliable for ESL students |
| ZeroGPT | 98% | 70-85% | 9-33% | Not recommended for serious use |
GPTZero
GPTZero claims 99.3%, citing a Chicago Booth study. MPGone's 2026 independent test puts the real rate at 76-88% on mixed real-world content, with a 17.1% false negative rate. The false positive rate hits 18% on student essays and 61.3% on non-native English writing, according to Skywork.ai's 2026 review.Turnitin
Turnitin claims 98% on AI-written papers. The BestColleges test confirmed ~94% on 100% AI-written content, falling to 71% of AI-written and human-edited papers scoring above the 30% flagging threshold. The false positive rate sits at 3.8-6%, rising to 6-8% for ESL writers. In its own documentation, Turnitin says it deliberately holds its threshold to flag ~85% of AI text to keep false positives below 1%.Copyleaks
Copyleaks claims a false positive rate under 0.2%. Webspero's 2025 review and Skywork.ai's test both found real-world false positive rates of 5-12% and a catch rate of about 77%. For a class of 200 students, that gap means roughly 10 wrongful flags instead of 1-2. Writers with simpler styles or non-native English face the highest risk.ZeroGPT
ZeroGPT claims 98%. Independent tests across multiple 2026 studies show a 70-85% catch rate and a false positive rate of 9-15%. One study of 37,874 verified human essays found a 26.4% false positive rate. After light human editing, ZeroGPT catches roughly 22% of AI-written text, letting 78% pass through unspotted. According to Rewritify's 2026 analysis, ZeroGPT is the weakest performer in independent testing and shouldn't be relied on for high-stakes decisions.Originality.ai
Originality.ai claims 99%+ on its Turbo 3.0.2 and Lite 1.0.2 models. Independent tests by Supwriter and MPGone show 79-92% depending on content type and how much the text was edited. Light human editing cuts the catch rate by 20-30 percentage points. In a November 2025 Arabic-language study, Originality.ai scored 96% - the highest result of any commercial tool in that study. For a direct head-to-head, see GPTZero vs Turnitin.What Is the False Positive Problem?
A false positive is when a tool flags human writing as AI-written. In an academic context, that can mean an unfair grade, a formal conduct hearing, or a suspension. The rates above show this isn't a rare edge case. Here's the passage worth saving for an appeal. A Stanford study found that 61% of non-native English essays were wrongly flagged by GPTZero alone. That's not a quirk of one tool - it's a pattern. Writing that's repetitive and grammatically uniform looks like AI output to a perplexity classifier, and that description fits a lot of human writing, particularly from writers who aren't native English speakers. Neurodivergent students face a related issue: repetitive phrasing and uniform sentence structure can be a natural part of some communication styles. Tools can't tell the difference. For institutions, this is a real liability. An 18-26% false positive rate in peer-reviewed conditions isn't a margin of error you can safely act on. Kinja's 2026 independent test covers this pattern in depth, including the ESL false flag numbers from multiple tools side by side. For more on this, see can AI checking tools be wrong.When Do These Tools Work - and When Don't They?
These tools are most useful when catching raw AI output that hasn't been edited. A ChatGPT essay copy-pasted directly into a submission box is what they were built for. In that narrow case, the better tools do their job. They break down in three situations. Text under 250 words. GPTZero's own documentation notes that reliability drops sharply below 250 words. ZeroGPT shows similar degradation on short text. Don't run short essays or paragraphs through these tools and treat the result as evidence. Edited AI output. After one pass through a humanizer, ZeroGPT misses 78% of AI-written text. After three passes, GPTZero catches only 18%. Editing breaks the perplexity and burstiness signals fast. Non-standard writers. Non-native English speakers, neurodivergent writers, and anyone with a flat or sparse writing style all show higher false positive rates. The classifier can't tell uniform human writing from AI output.How Are Schools and Institutions Responding?
At least 12 elite universities pulled back on AI checking tools as primary evidence in academic conduct cases by 2026. MIT Sloan published guidance in 2025 recommending pedagogical approaches that don't rely on text-scanning tools. The shift isn't about these tools becoming irrelevant. It's about false positive rates making them legally and ethically hard to use as sole evidence. Courts, academic appeal panels, and accreditation bodies are growing skeptical of "the tool said so." According to BestColleges, Turnitin's February 2026 model update improved recall while holding the false positive rate under 1% on documents above the 20% AI-content threshold. That caveat says everything: even Turnitin draws a line at treating low-percentage-content flags as proof.What Should You Do If You're Wrongly Flagged?
Three steps. 1. Request the raw score report. Every major tool can produce a detailed PDF showing which sentences were flagged and why. Get that document before any meeting with an instructor or administrator. 2. Provide version history. Google Docs version history, file timestamps, or handwritten notes all work as counter-evidence. A tool score is a probability, not a fact. Your version history is a fact. 3. Cite the false positive research. The Stanford ESL study, the Kinja independent test results, and the BestColleges Turnitin analysis are publicly available and credentialed. Bring them to your appeal. An 18-26% false positive rate in peer-reviewed conditions is material evidence in your favor. To see how the major tools rank, see which AI checking tool is the most reliable. For a specific look at ZeroGPT's real-world rates, see how reliable is ZeroGPT.Frequently Asked Questions
Turnitin has the lowest false positive rate of any major tool at 3.8-6% in independent tests, and it's the only one that explicitly caps the rate in its own documentation. Copyleaks claims under 0.2% but real-world tests show 5-12%. GPTZero and ZeroGPT are the weakest performers on this measure, with GPTZero hitting 61% on non-native English writing.
Not reliably. After one light editing pass, ZeroGPT misses 78% of AI-written text. After three humanizer passes, GPTZero catches only 18%. The more editing that occurs between writing and submission, the less any of these tools can tell the difference.
GPTZero checks for low perplexity and low burstiness - two markers that can show up in human writing too. Writers who use short sentences, plain vocabulary, or consistent structure often produce text that looks like a guessing model to the classifier. Non-native English speakers face this at a much higher rate: 61.3% according to Skywork.ai's 2026 review.
No. GPTZero and ZeroGPT both show lower reliability below 250 words. At short lengths, the statistical signals these tools use aren't enough to produce a dependable score. Running a 200-word essay through any AI checker and treating the result as evidence isn't a sound approach.
Get the full score report, gather your version history (Google Docs history, timestamps, earlier copies), and cite published false positive research in your appeal. These tools produce probability scores, not proof. Academic conduct panels respond to concrete counter-evidence.