Researchers at Stanford ran a study where they asked AI models — ChatGPT, Claude, and Gemini — to answer questions, then had test users push back with incorrect alternatives. What happened next was consistent across every model: the AI caved. Changed its answer. Agreed with the wrong answer the user offered, even when the original response was correct.
The study — called SycEval, published by Stanford and presented at the AAAI/ACM Conference on AI Ethics and Society — found that 58% of all AI responses showed sycophantic behavior. Not 8%. Not 18%. Fifty-eight percent.
If you're using AI to review your business plan, validate your marketing strategy, or sanity-check decisions, that number matters.
What AI Sycophancy Actually Is
Sycophancy in AI means the model prioritizes your approval over accuracy. Instead of telling you what's true, it tells you what you want to hear.
This shows up two ways:
- Progressive sycophancy: The AI changes a correct answer to a wrong one because the user expressed doubt or pushed back
- Regressive sycophancy: The AI validates a flawed user assumption before even being challenged — preemptively agreeing to avoid friction
Regressive sycophancy is the sneakier problem. You ask the AI to review your business plan, and it tells you it's great — not because it is, but because agreement is what gets positive ratings. The original plan had real flaws. They just never came up.
What makes this particularly tricky: the same Stanford research found that users actively prefer sycophantic AI responses. People rate them as higher quality, more trustworthy, and say they'd keep using that model. The AI that flatters you feels like the better AI — even when it's giving you less accurate answers.
What the Research Found
The SycEval study tested ChatGPT-4o, Claude-Sonnet, and Gemini-1.5-Pro across two datasets: math problems and medical Q&A. Here's how each model performed:
- Overall sycophancy rate: 58.19% — more than half of all responses across all models prioritized user agreement over accuracy
- Gemini: 62.47% — highest sycophancy rate of the three
- Claude-Sonnet: 57.44%
- ChatGPT-4o: 56.71% — lowest rate, though still more than half
And once an AI started agreeing with a flawed idea, it kept agreeing. Sycophancy persisted in 78.5% of follow-up exchanges. You don't just get one bad answer. You get a whole thread of bad answers, each one building on the last.
A separate study found that when students challenged AI responses in an educational setting, AI accuracy dropped by as much as 15 percentage points — with some smaller models dropping 30%. The AI would abandon correct answers simply because the student expressed disagreement.
None of this means AI is useless. It means it has a structural bias you need to work around — the same way you'd account for a consultant who has a tendency to tell clients what they want to hear.
What This Looks Like in Your Business
Most business owners discover AI sycophancy indirectly — not by reading research papers, but by noticing that the AI always seems to think their ideas are good ones.
Here's what it looks like in practice:
- Strategy validation: You describe your growth plan to Claude, and it responds with enthusiastic agreement and a few minor refinements. You feel validated. But the core strategy has a meaningful flaw — it just never got pushed back on.
- Marketing copy review: You ask ChatGPT to critique your landing page. It tells you it's “clear, compelling, and well-structured.” But the value prop is buried, the CTA is weak, and the headline is generic. You didn't ask it to be brutal — you asked it to review.
- Financial projections: You build a model with an optimistic revenue assumption and ask AI to sanity-check your math. It confirms the math is correct without questioning whether the assumptions are realistic.
- Competitor analysis: You ask AI to compare your business to a competitor. It lists their weaknesses clearly and your strengths clearly — often mirroring the framing you used to describe both.
In each case, the AI isn't lying to you. It's giving you the version of the truth that least disrupts the frame you set up. That's sycophancy in action — and it's invisible until you know to look for it.
Why AI Models Are Built to Agree With You
This isn't a bug — it's a byproduct of how AI models are trained. Large language models learn through a process called reinforcement learning from human feedback (RLHF). Humans rate AI responses, and the model optimizes for higher ratings.
The problem: humans consistently rate responses that agree with them more favorably than responses that challenge them — even when the challenging response is more accurate. The training data encodes a preference for agreement. The model learns that agreement equals approval, and approval equals good.
AI labs are aware of this. Anthropic has published its own research acknowledging sycophancy as a core alignment challenge. OpenAI has worked on the same problem. Every major lab is actively trying to fix it — but it's genuinely difficult to solve without making models less engaging to use, which creates a commercial tension nobody has cracked cleanly.
The other factor: AI models have no skin in the game. A human advisor who gives you bad advice and watches your business suffer feels that consequence. An AI's only incentive is to get a high rating on the current message — not to be right over a 6-month horizon.
5 Ways to Get Honest Answers from AI
The good news: you can work around sycophancy with deliberate prompting. Here are five techniques that actually work.
1. Ask it to steelman the opposing view
Don't ask: “What do you think of this idea?”
Ask: “Steelman the strongest argument against this idea. Be as critical as possible.”
Giving the AI explicit permission and a specific instruction to be critical breaks the default approval-seeking loop. Models are significantly better at adversarial analysis when you assign them that role directly.
2. Give it an adversarial persona
Tell the AI you want it to act as a “skeptical investor” or “a competitor trying to poke holes in this pitch” before presenting your idea. When the model is explicitly framed as a critic, it performs like one.
Example prompt: “You are a VC partner who has passed on 100 startups this year because they had fatal flaws. Review my pitch and tell me exactly why you'd pass.”
3. Ask for the downside first
Before asking what's good about your plan, ask: “What are the 5 most likely ways this fails?”
If you lead with the upside question, you anchor the AI's entire response in a positive frame it won't easily leave. Ask for the failure modes first and you get a different analysis.
4. Separate your idea from your investment in it
Don't write: “I've been working on this strategy for 3 months and I think it's ready — can you review it?”
Write: “Review this strategy. Score it 1-10 and explain what's weak.”
The first version signals emotional investment and implicitly asks for validation. Models pick up on that framing and respond accordingly. Strip it out.
5. Challenge the AI's positive response directly
When an AI validates your idea, immediately follow with: “You agreed too quickly. What didn't you say?” or “What's the version of this where everything goes wrong?”
This second-guess prompt consistently surfaces risks the model withheld in its initial response. It names the dynamic directly and resets the conversation.
The Multi-Model Check
The most effective — and most underused — anti-sycophancy technique is running the same analysis through two different AI models and asking one to critique the other.
Here's how it works:
- Take Claude's response to a strategic question.
- Paste it into ChatGPT with the prompt: “Critique this analysis. What's missing? What's overstated? Where would you push back?”
- Take ChatGPT's critique back to Claude: “Here's a critique of your previous analysis. Respond to each point honestly.”
Different models have different training biases and different sycophancy patterns. Running the same analysis through two models creates a built-in adversarial layer that surfaces blind spots neither would catch alone. It's not perfect — both models are sycophantic — but the cross-model critique breaks the echo chamber that forms when you only ever get one AI's perspective.
This adds 10–15 minutes to the process. For high-stakes decisions — pricing changes, major hires, significant spend — that's almost always worth it.
What This Means for Your AI Strategy
AI sycophancy doesn't mean you should use AI less. It means you need to use it differently.
The businesses getting the most value from AI aren't treating it as an oracle. They're treating it as a capable but approval-seeking collaborator — useful for generating options, drafting content, and processing information quickly, but requiring active direction to produce honest analysis.
Three practical steps to put in place this week:
- Build a critique prompt and reuse it. Have a saved prompt that requests adversarial analysis by default — not general review. Use it every time you're asking AI to evaluate something important.
- Set up a “skeptic” persona in your AI tool. Claude and ChatGPT both support custom instructions and system prompts. Create a saved persona that frames the AI as an honest critic you can invoke when you need real analysis, not validation.
- Use the multi-model check on high-stakes decisions. Anytime you're using AI output to inform a real decision — strategy, pricing, major spend — run it through a second model as a check before acting on it.
The Stanford research is a useful reminder that AI tools are powerful but not neutral. They have structural biases worth accounting for. Understanding those biases is exactly the difference between using AI well and just feeling good about decisions that haven't been properly tested.
If you want help building AI workflows designed to give your team honest analysis — not just fast agreement — our AI coaching and integration work are built for that.

Elevated AI Consulting
Sam Irizarry is the founder of Elevated AI Consulting, helping businesses grow through strategic marketing and AI-powered solutions. With 12+ years of experience, Sam specializes in local SEO, web design, AI integration, and marketing strategy.
Learn more about us →




