Treat AI output as a draft.
This sounds obvious. But watch how people actually use LLMs: copy, paste, send. No review. Full trust.
The fix isn't complex. It's a 30-second check before anything goes out. Here's the checklist I use.
The 30-Second Checklist
Before you use any AI-generated content:
1. Factual Claims (5 seconds)
- Are there specific numbers, dates, or statistics?
- Are there cited sources?
- Can I verify the important ones?
If yes, spot-check at least one. LLMs confabulate citations confidently.
2. Tone Match (5 seconds)
- Does this sound like me/my organization?
- Is the formality level right?
- Any phrases that feel off?
If something feels "too AI," it probably is.
3. Logical Coherence (10 seconds)
- Does the conclusion follow from the premise?
- Are there contradictions?
- Does it actually answer what I asked?
Skim for the argument structure, not just the words.
4. Risk Assessment (10 seconds)
- What happens if this is wrong?
- Who sees this output?
- What's the cost of an error?
Higher stakes = more careful review.
When to Skip (Carefully)
Not everything needs the same scrutiny:
Low-stakes: Brainstorming ideas, first drafts, personal notes
- Quick skim is fine
Medium-stakes: Emails, documentation, internal reports
- Full checklist
High-stakes: Client deliverables, public content, code in production
- Full checklist + domain expert review
What I've Caught
Using this checklist, I've caught:
- Fake citations - Papers that don't exist, misattributed quotes
- Subtle logic errors - Conclusions that don't follow from evidence
- Wrong numbers - Statistics that were plausible but invented
- Tone drift - Responses that didn't match my voice
- Hallucinated features - Code examples using APIs that don't exist
Every one of these would have been embarrassing (or worse) if published.
The Meta-Point
This isn't about distrusting AI. It's about avoiding automation bias - the human habit of over-trusting "the system" when it sounds confident.
LLMs are useful. But they don't own the consequences. You do.
So: draft first, verify second, ship third.
Making It Automatic
The checklist becomes habit fast. After a week, you'll do it unconsciously:
- Pause before copying
- Scan for claims
- Check the tone
- Assess the stakes
- Then use it (or don't)
Five seconds most of the time. Thirty seconds when it matters. Zero embarrassing emails sent.
For Teams
If you're implementing AI tools in a team:
- Document the checklist - Make it explicit
- Share failure examples - Show what gets caught
- Normalize "I verified this" - Make checking the expectation
- Build review into workflows - Don't rely on individual discipline
The goal isn't to slow people down. It's to make checking effortless.
My Take
I think about this like driving. You check your mirrors without conscious thought. It takes zero extra time because it's automatic.
AI review should be the same. Not a burden - a habit. Not paranoia - professionalism.
The people who use AI best aren't the ones who trust it most. They're the ones who verify instinctively.
Quick Reference
Before using AI output:
[ ] Factual claims verified?
[ ] Tone matches context?
[ ] Logic coherent?
[ ] Stakes assessed?
If high-stakes: get a second pair of eyes.
Print this. Stick it next to your monitor. Use it until you don't need it.
Further Reading
- NIST AI Risk Management Framework - Governance and human oversight framing
- Automation Bias and Accountability - Why friction reduces over-trust (Skitka et al.)
- Fabricated References in ChatGPT-Generated Content - High rates of inaccurate citations
Leave a Comment
Comments (0)
Be the first to comment on this post.
Comments are approved automatically.