TL;DR

Agents today are basically just chatbots with a to-do list. Real agents (the kind that reliably plan, act, and execute) need structure, constraints, and continuous evaluation. The recipe is to: write explicit specs → separate planning and execution → ground every action → verify outputs → add human-in-loop for uncertainty. Don’t treat your agent like a science experiment. You have to treat them like production software.

A quick note

Amazon just announced 30,000 layoffs. UPS cut 48,000. Both cited AI as a key driver, and it’s not stopping there.

If you were recently affected and want to pivot into an AI-related role, I want to help. I’m putting together a small effort to connect people with resources, mentors, and companies hiring for AI-skilled roles.

I’ve been in the AI space for 8+ years, worked with ML systems at Meta, founded an AI education non-profit that reached 70,000 people, and now run an AI testing platform where I see firsthand how companies are implementing AI and reshaping their approach to business.

If that sounds useful, you can fill out the form below. I’ll share what I learn as I help people navigate this shift.

Why “good” agents are rare

Building an AI agent isn’t about letting a model “figure things out.” It’s about engineering a controlled system where the model can plan, act, and learn — safely.

Agents fail for predictable reasons:

  • Ambiguous goals. The model doesn’t know what success looks like.

  • Unbounded tools. It can call anything, in any order.

  • No feedback loops. Mistakes compound without correction.

  • No monitoring. You don’t know when (or why) it broke.

What you get: hallucinated actions, endless loops, or confident wrong behavior.

What you want: a system that can reason, act, verify, and recover all under supervision.

The agent stack (layered like any production system)

1. Spec first — write the job description

Before any prompt or code, define:

  • Goal: what outcome is the agent responsible for?

  • Allowed tools: which APIs or actions can it use?

  • Constraints: what it can’t do (e.g., “never approve payments > $1k”).

  • Evidence requirements: every action must cite where it got the data.

This becomes your agent spec. Basically, the contract between you and the system.

2. Plan, then execute (never both at once)

Separate planning (thinking) from doing (acting).

Example:
1️⃣ Planner: “I’ll fetch the ledger, check balances, and draft a transfer.”
2️⃣ Executor: runs those steps deterministically through verified APIs.

This separation gives you reproducibility, auditability, and safety.

3. Ground every action in truth

Agents hallucinate just like models do — they make up actions when context is missing.
So ground everything:

  • Connect to deterministic APIs for facts (ledger, CRM, docs).

  • Force references: every action output must include evidence_refs.

  • Ban open-ended calls like “search the web and decide.”

If it can’t verify, it shouldn’t act.

4. Verify before you trust

Use structured outputs and post-checks:

  • Schema validation (does it match the spec?)

  • Sanity checks (does the number make sense?)

  • Approval gates (manual or rule-based for sensitive actions)

Verification slows you down slightly, but it prevents silent failure loops.

5. Evals and monitoring (the secret sauce)

Agents evolve over time, so evals must, too.
Build evals that measure:

  • Task completion rate — % of plans that finish successfully

  • Safety violations — unauthorized actions or skipped verifications

  • Fallback rate — % routed to human review

  • Mean time to recovery — how fast failures are caught and corrected

Run replay evals on production logs weekly. Agents learn; your evals should, too.

6. Human in the loop (the ultimate safety rail)

Don’t aim for full autonomy first.
Route low-confidence or high-risk tasks to humans — ideally with AI drafting the action for approval.
You’ll get reliability and user trust.

Small example — what good looks like

A “Meeting Notes Agent” that:

  1. Fetches meeting transcript from your DB

  2. Plans a 3-step summary → action list → follow-up email

  3. Executes only if all referenced data is verified

  4. Sends the draft for approval

  5. Logs every step (inputs, outputs, evidence)

That’s an agent you can ship without fear.

Quick checklist

  • Define a spec: goal, tools, constraints, evidence.

  • Split planner vs executor.

  • Add schema + output validation.

  • Require evidence for every fact.

  • Route low-confidence steps to humans.

  • Add daily evals + production logging.

  • Monitor safety and success metrics.

Final note

Agents aren’t dumb, all they need is some structure and constraints. Structure creates reliability. Constraints create trust. And that’s how you get a good agent.

Agents should be boringly predictable before they’re impressively autonomous.
Start simple, measure everything, and evolve with evals.

👉 If you found this issue useful, share it with a teammate or founder navigating AI adoption.

And subscribe to AI Ready for weekly lessons on how leaders are making AI real at scale.

Until next time,
Haroon

Reply

or to participate

Keep Reading

No posts found