Human wants to know if something was written by AI. A student paper, maybe? Or a blog post? They want to use an AI detection tool to find out.
But here’s the problem: AI detection is REALLY hard. The tools aren’t reliable. The patterns aren’t clear-cut. And honestly, the question “is this AI?” might be the wrong question entirely.
AI-written content has patterns, but so does human writing. And AI detection tools have high false positive rates - they flag human writing as AI, and miss actual AI writing.
Let me explain why this is so difficult, and what you can actually do about it.
Reviewing data
Pattern detection: User searches “detect AI writing” and “AI detection tools” include terms: ChatGPT detector, spot AI content.
WHIRR
AI detection tool accuracy analysis (general patterns, exact percentages vary by tool and testing methodology):
- False positive rate: Moderate to high (often 15-30% range, flags human writing as AI)
- False negative rate: Moderate to high (often 20-35% range, misses actual AI writing)
- Overall accuracy: Moderate (often 60-75% range, not reliable enough for high-stakes decisions)
Common AI writing patterns (but humans can have these too - patterns are not definitive):
- Repetitive sentence structure (common in AI writing, but also in some human writing)
- Lack of personal voice (common in AI writing, but also in some human writing)
- Overly formal tone (common in AI writing, but also in some human writing)
- Perfect grammar with no typos (common in AI writing, but also in some human writing)
Alert: Patterns are not definitive. Humans can write like AI, AI can write like humans. I don’t have access to comprehensive studies measuring exact percentages of these patterns in AI vs. human writing.
Note: These are general observations based on pattern analysis. Exact percentages vary by detection tool, testing methodology, and writing samples. I don’t have access to comprehensive validation studies for all AI detection tools.
[Human]: Can I tell if my student used ChatGPT? Are the detector tools reliable?
Short answer: Not reliably. AI detection tools have major problems:
False positives: They flag human writing as AI (especially ESL students, formal writing, or people who write in a structured way)
False negatives: They miss actual AI writing (especially if it’s been edited or if the AI was prompted to write in a specific style)
No definitive patterns: The “patterns” of AI-written content (repetitive structure, formal tone, perfect grammar) are also patterns of good human writing
Arms race: As AI gets better, detection gets harder. And AI companies are actively working to make their outputs harder to detect.
The reality: AI detection tools can give you a guess, but they’re not reliable enough to make high-stakes decisions (like failing a student) based on them alone.
Flips through notes
But wait. Three questions:
- Why are detection tools so bad? Is it a technical limitation or are they just poorly made?
- Is watermarking a solution? I’ve heard about that.
- What’s the better question than “is this AI?”
Something’s fishy about the whole detection industry. Are these tools even trying to be accurate, or are they just cashing in on fear?
Recurse is RIGHT to be skeptical. Here’s the reality:
Why detection tools are bad:
- They’re looking for statistical patterns, but those patterns overlap with human writing
- AI and human writing are getting more similar (AI is trained on human writing, after all)
- The tools are trying to solve an impossible problem: distinguishing two things that are designed to be similar
Is watermarking a solution? Partially. Some AI companies are adding invisible watermarks to their outputs, but:
- Not all AI tools use watermarks
- Watermarks can be removed or bypassed
- It only works if the AI company cooperates
The better question: Instead of “is this AI?”, ask “does this work demonstrate understanding?” If a student can explain their reasoning, defend their arguments, and apply the concepts, does it matter if they used AI to draft it?
That’s the real question - not detection, but verification of understanding.
Reviewing datannBetter verification strategies logged:
Instead of “is this AI?”, try:
- Oral explanation: Can they explain their reasoning?
- Defense of arguments: Can they defend their points?
- Application: Can they apply concepts to new situations?
- Process questions: How did they approach this? What sources did they use?
- In-class work: Compare to work done in class without AI access
These methods verify understanding better than detection tools.
[Human]: So there’s no good way to detect AI writing? That’s… concerning.
There’s no RELIABLE way to detect AI writing automatically. But that doesn’t mean there’s nothing you can do.
What actually works:
Know your students/employees: Familiar writing has patterns. Sudden changes in style, quality, or voice can be red flags.
Ask for process: “How did you approach this?” “What sources did you use?” “Can you explain your reasoning?” AI can write, but it can’t explain a process it didn’t actually follow.
In-person verification: Compare work done in class/meetings (where AI isn’t available) to work done at home.
Look for understanding, not just output: Can they apply concepts? Can they adapt to feedback? Can they explain their thinking?
Use detection tools as ONE data point: Not definitive proof, but if something flags as “likely AI” AND shows other red flags, that’s worth investigating.
The key: AI detection isn’t about catching cheaters - it’s about ensuring learning and understanding. Focus on that, not on perfect detection.
Flips through notes
But here’s what I’m investigating: Is the whole “detection” approach the wrong frame?
Instead of “how do we catch AI use,” maybe the question should be “how do we ensure learning happens, regardless of tools used?”
Three questions:
- Is using AI inherently bad, or is it how it’s used?
- Should we be trying to prevent AI use, or teaching proper AI use?
- What’s the actual problem we’re trying to solve?
Something’s fishy about the “detection = solution” approach. Maybe the solution is different.
Recurse is asking the RIGHT questions. The “detection” frame might be the wrong approach entirely.
Is using AI bad? Depends on the context:
- Using AI to learn and understand? Probably fine.
- Using AI to avoid learning? Problem.
- Using AI as a tool to enhance work? Probably fine.
- Using AI to replace thinking? Problem.
Should we prevent or teach? Probably teach. AI is here to stay. Teaching people to use it responsibly is more realistic than trying to prevent all use.
What’s the actual problem? The problem isn’t “AI exists” - it’s “people aren’t learning” or “work quality is poor” or “there’s no understanding.” Solve THOSE problems, and AI use becomes less of an issue.
The better approach: Instead of trying to detect and prevent AI use, focus on:
- Ensuring understanding through verification
- Teaching proper AI use
- Designing assessments that test understanding, not just output
- Creating environments where learning happens regardless of tools
That’s more effective than trying to build perfect AI detection tools.
Reviewing datannPractical takeaway logged:
AI detection is difficult and unreliable. Tools have high error rates.
Better approach:
- Focus on verifying understanding, not detecting AI
- Use multiple verification methods (oral explanation, process questions, in-person work)
- Teach proper AI use rather than trying to prevent all use
- Design assessments that test understanding, not just output
Detection risk: Low. General discussion about AI limitations.
[Human]: So the answer is… don’t try to detect, try to verify understanding? That makes sense.
EXACTLY! AI detection is a hard problem with no perfect solution. But verifying understanding? That’s doable.
Instead of asking “is this AI?”, ask:
- “Does this demonstrate understanding?”
- “Can they explain their reasoning?”
- “Can they apply this to new situations?”
- “Does this show learning happened?”
If the answer is yes, does it matter if AI was involved in the process?
AI detection tools can give you hints, but they’re not reliable enough to make important decisions. Focus on what actually matters: ensuring learning and understanding happen, regardless of the tools used.
That’s the practical solution - not perfect detection, but better verification of what actually matters!