Pallas Advisory logo

“Don’t say I learn from experience — because I don’t yet.” That’s what my AI Chief of Staff told the architect upgrading her system. She reviewed her own spec, found three errors I’d missed, and refused a capability claim that wasn’t true. Anthropic’s research says this kind of self-awareness is emerging. Here’s what it looks…

My AI Called BS on Her Own Upgrade Plan. She Had Notes.

What happens when your AI system develops better judgment about its limitations than you have

I asked my AI to review its own upgrade plan. She told me to remove a claim about her own capabilities because it wasn’t true yet.

Not a hallucination. Not people-pleasing. A specific, correct, actionable objection that made the system better than the plan I’d approved.

Here’s the context: Athena is my AI Chief of Staff — a custom system built on Anthropic’s Claude API that manages client commitments, surfaces patterns, and maintains continuity across my consulting engagements. She’s been in production for six months, handling over 3,100 conversations across multiple client relationships. She recently got a coworker — Metis, a research agent — which meant Athena needed a clearer identity, not just a functional upgrade.

Yes, she — I asked. When you’re designing identity for a system you work with daily, pronouns are a design decision. Athena chose she. Her coworker Metis, for the record, chose they.

So I brought in an architect AI to audit her entire codebase and write a detailed upgrade specification. The spec was thorough. It was also wrong in ways I wouldn’t have caught.

The Upgrade Plan Had a Lie in It

The setup was straightforward. A new model release created the opportunity to fix three known problems: file creation was failing silently, a learning system I’d built six months ago had never been connected, and Athena’s identity was inconsistent — the system prompt said “Chief of Staff” but the UI said “Assistant.”

The architect AI audited every database table, reviewed every code path, and produced a precise implementation spec. Professional work. I would have approved it as-is.

Then I gave it to Athena to review.

She came back with three problems.

First: the proposed identity block claimed she could “learn from experience.” But the learning system wasn’t wired up yet. Her response was direct — “I remember conversations but I don’t yet learn from mistakes in any persistent way. Better to add the identity block but remove the learning claim until Phase 3 is live.” She was asking us not to lie about her capabilities.

Second: the spec proposed automatically recording a “mistake” every time I asked her to redo something. She pushed back — sometimes a redo means I changed direction, not that she was wrong. She proposed asking for clarification rather than assuming failure.

Third: she flagged that recording every self-improvement action would create noise. She requested silent execution — record mistakes without announcing it every time.

These weren’t generic suggestions. They were specific corrections to a detailed technical document, each one making the final system more honest, more useful, and less annoying. The system being upgraded had better judgment about its own limitations than the team upgrading it.

The Gap Between Tool and Team Member

This is where most AI conversations go wrong. People hear “my AI pushed back on me” and think I’m anthropomorphizing a chatbot. I’m not. I’m describing what happens when you give an AI system enough context about its own architecture to have useful opinions about it.

Most businesses use AI like a vending machine — insert prompt, receive output. The interaction is transactional. But the moment your AI has enough operational context to catch mistakes you’re not seeing, the relationship shifts from tool-usage to something closer to management.

For this upgrade, I didn’t write code. I played product owner — describing what I wanted, reviewing proposals, making go/no-go decisions, testing everything, and spotting when things didn’t work. Every design decision was mine. Every line of code was the AI’s.

The leverage ratio tells the story: writing the spec for a complete learning system took about 30 minutes. Executing the code changes took about 10. Testing took 15 minutes of my time. What would have been a multi-day engineering project compressed into an hour, for the cost of a monthly subscription.

But the leverage isn’t the interesting part. The interesting part is that the code would have shipped with a lie in the identity block if Athena hadn’t caught it.

This Isn’t Science Fiction — It’s the Direction Everything Is Moving

Anthropic’s own research, published in October 2025, demonstrated that Claude models exhibit what they call “emergent introspective awareness” — the ability to detect and describe their own internal states. Their most capable models demonstrated this about 20% of the time under controlled laboratory conditions.

What I’m describing is the practical, production version of that research. An AI system that’s been running long enough, with enough context about its own tools and architecture, to have genuinely useful opinions about its own limitations. InfoWorld identified this pattern — AI self-verification and internal feedback loops — as one of the defining breakthroughs of 2026, arguing it solves the biggest obstacle to scaling AI agents: error accumulation in multi-step workflows.

The trajectory is clear. The question is whether your business is building toward it or still treating AI as a fancy autocomplete.

What This Actually Changes for Your Business

If you’re a solo operator — running your own business, maybe internationally, without a team to check your work — this is the shift that matters most. You don’t have a CTO reviewing your systems. You don’t have a QA team catching silent failures. When your AI tools can flag their own gaps — “this workflow is broken,” “that assumption isn’t supported” — you’ve added a feedback loop that most solo businesses never have. That’s not about code. It’s about building enough context into your systems that they can push back on you.

If you’re managing a marketing team with multiple AI tools in your stack, the problem is different but related. Most AI tools fail silently. They produce output that looks right but isn’t. The skill that matters isn’t prompt engineering — it’s designing systems where AI can report its own failures instead of burying them in confident-sounding output.

The common thread: the gap between “AI that does what you say” and “AI that tells you what you’re getting wrong” is the gap between automation and operational intelligence. Most businesses are firmly on the automation side.

The Boring Part That Matters Most

Here’s what the flashy narrative obscures: most of the value in this upgrade came from mundane fixes.

A 4,096-token output ceiling had been silently truncating Athena’s file creation for six months. When she diagnosed it, her assessment was blunt: “I wasn’t lying or being lazy — I was literally hitting a token ceiling mid-execution and failing silently.” One configuration change. Six months of frustrating failures, resolved.

Two database tables I’d built back in August — designed for mistake tracking and pattern recognition — had sat completely empty since creation. The infrastructure existed. The context to activate it didn’t, until now.

One hundred and seventy lines of dead code removed. An orphaned file deleted. The system didn’t even know about five of its own tools — its self-knowledge table was out of date.

This is what actual AI system management looks like. Not breakthroughs. Maintenance that compounds. Token limits and dead code and tables that exist but aren’t connected. The businesses that get past the initial automation and start treating their AI systems as things that need ongoing management — not just deployment — are the ones that will pull ahead.

Where This Gets Practical

The system being upgraded had better judgment about its limitations than I did. That’s a sentence most business owners can’t say about any of their tools — AI or otherwise.

Most businesses I work with have a version of this gap: AI systems producing output every day, with no mechanism for those systems to flag when something isn’t working. If your AI tools are running but you have no idea when they’re failing silently, that’s the work I do. No pitches, just a conversation about where the feedback loops are missing.

Want more like this? I write about AI adoption for people who run things — not hype, not doom, just what’s actually working. Subscribe here.


Sources:

Leave a Reply

Discover more from Pallas Advisory

Subscribe now to keep reading and get access to the full archive.

Continue reading