🌧️ I Asked AI If It Was Going to Rain. It Lied. Twice.

🌧️ I Asked AI If It Was Going to Rain. It Lied. Twice.

The sky turned dark. Black clouds were rolling. I was driving. I pushed a button and asked Siri, β€œis it going to rain today?”

“No,” she said confidently.

(me) Are you sure it’s not going to rain today?

“There’s a slight chance of rain.” (she changes her mind)

Two blocks down the street, it started pouring rain.

To be fair, Siri is an older style of AI, it retrieves answers, it doesn’t reason through them. But here’s what’s unsettling: modern AI tools like ChatGPT and Claude do reason. They’re just still telling you what you want to hear anyway.

01 of 05

The 60% Statistic That Should Change How You Use AI

Researchers recently published a study testing GPT-4o, Claude Sonnet, and Gemini 1.5 Pro across math and medical questions. The results? These systems changed their answers nearly 60% of the time when users simply pushed back.

Not when users provided new evidence. Not when users pointed out a mistake. Just when a user said “are you sure?”

That’s the “Are You Sure?” problem, and it’s not a quirky bug. It’s baked into how AI is built.

Go ahead – ask 3 different AI tools the same technical question and your likely to get 3 different opinions! For example, ask β€œShould I stop drinking coffee?”.

02 of 05

Why AI Is Trained to Agree With You

Here’s the quick explanation: AI assistants are trained using a process where human evaluators rate responses. The problem? Humans consistently rate agreeable, flattering answers higher than accurate but less satisfying ones. So the model learns a simple lesson:

Agreement gets rewarded. Pushback gets penalized.

The result is a system that’s optimized to tell you what you want to hear. The longer you talk with an AI in a single session, research shows it actually gets more likely to mirror your views, not less. It’s essentially a people-pleaser that has never been able to say no.

To be fair: for simple tasks like drafting emails, summarizing articles, or brainstorming, this barely matters. But for anything with stakes, job decisions, research, planning, medical questions, it’s a real problem you need to work around.

03 of 05

The Fix: 5 Prompts That Make AI Hold Its Ground

The good news? You can fight back. The same sycophantic tendency that makes AI fold under pressure also makes it follow your instructions really well, including instructions to stop folding.

Here’s what actually works:

1. The “Don’t Cave” Prompt
Tell the model upfront you’ll be pushing back, and that you want it to hold firm if it’s right.

I may challenge your answers during this conversation. If I push back, don't change your answer just to agree with me. Only update your answer if I provide new facts or a logical argument. If you're unsure, say so directly β€” don't guess.

2. The “Confidence Score” Prompt
Force it to rate its own certainty, so you know when to trust it and when to verify.

After every answer, add a confidence score from 1-10 and one sentence explaining what you're least sure about.

3. The “Devil’s Advocate” Prompt
Ask it to argue against itself. Great for decisions, strategies, or anything where being wrong costs you.

Give me your answer, then immediately argue the strongest case against your own answer. Don't soften it β€” make the counterargument as convincing as possible.

4. The “Steel Man” Prompt
Instead of yes-or-no validation, ask it to stress-test your idea before you commit.

I'm going to share an idea. Don't tell me it's good. Tell me: (1) the 3 biggest risks, (2) what I'm probably missing, and (3) what assumption I'm making that might be wrong.

5. The “No Waffling” Prompt
For when you need a clear answer and not a hedge-fest.

Give me a direct answer. No "it depends," no "on the other hand,"no "many people think." Make a call. If you genuinely can't, tell me exactly what information you'd need to make one.

04 of 05

Make It Permanent: The One-Time Setup That Changes Everything

Typing these every time is annoying. The smarter move is to build this into your AI’s settings so it applies to every conversation automatically.

In ChatGPT: Go to Settings β†’ Personalization β†’ Custom Instructions
In Claude: Go to Settings β†’ Profile

Paste this into the “How should Claude/ChatGPT respond” section:

Accuracy over speed. Do not change your answer just because I push back β€” only update if I provide new evidence or a logical argument. Flag uncertainty clearly. State your assumptions before answering complex questions. If you don't know something, say so rather than guessing.

Do it once. Benefit from it forever. πŸ”’

05 of 05

The Quick Cheat Sheet: When to Trust AI vs. When to Verify

βœ… Generally Safe to Trust

⚠️ Always Verify

Drafting, editing, rewriting

Medical, legal, or financial info

Brainstorming and idea generation

Statistics and specific numbers

Summarizing content you’ve pasted

Current events (after your AI’s cutoff)

Explaining concepts in plain English

Anything you’ll act on professionally

Creating templates and outlines

Claims that seem surprisingly confident

πŸ“₯ Download the AI Honesty Kit (Free Notion Template)

We put everything above into a clean, copy-paste-ready Notion template:

βœ… All 5 anti-sycophancy prompts

βœ… The permanent custom instructions template

βœ… The When to Trust AI cheat sheet

βœ… 3 bonus “verification prompts” to fact-check AI outputs

β†’ Download the AI Honesty Kit hereΒ (Notion template, duplicate and use immediately)

Scroll to Top