Skip to content
LEWIS C. LIN AMAZON.COM BESTSELLING AUTHOR
Go back

When to Use AI Agents

Edit page

When to Use AI Agents

Half the AI agent pitches I see are just expensive chatbots with delusions of grandeur. The other half are genuinely useful—but probably don’t need an agent at all.

Here’s the thing: agents aren’t just “better automation.” They’re fundamentally different beasts that excel at messy, unpredictable problems where the right response depends on context that’s constantly shifting. But if your workflow is mostly predictable? You’re probably better off with good old-fashioned automation.

So how do you tell the difference? I’ve been thinking through this problem and landed on a simple framework—not a scoring system or some pseudo-scientific methodology, but a way to think through whether your problem actually needs an agent.

What Actually Makes Something “Agentic”

Real agents do four things well: they observe their environment, reason through problems step-by-step, adapt when things change, and act with minimal babysitting. The key word is “minimal”—if you’re constantly course-correcting an “agent,” it’s probably just automation with extra steps.

The litmus test: could a competent human handle this task if they were dropped into the situation cold, with just the same information your system has access to? If the answer is “probably not without extensive training,” you might not need an agent—you need better processes or clearer requirements.

Five Ways to Think About Agent-Worthy Problems

Instead of scoring pillars, I find it helpful to ask five questions about any potential agent deployment:

1. Does the context actually change?

The question: How often do the relevant facts of your situation shift in ways that matter for decision-making?

Why it matters: This is the biggest separator between agent and automation territory. If your process works the same way regardless of external conditions, automation will be faster, cheaper, and more reliable.

Examples that need agents:

Examples that don’t:

Red flag: If you find yourself saying “it’s always the same except for…” you probably want automation with some conditional logic, not an agent.

2. Are there actually multiple steps that depend on each other?

The question: Does success require chaining together several decisions where the outcome of step 1 affects what you should do in step 2?

Why it matters: True multi-step reasoning is computationally expensive and error-prone. If your process is mostly parallel or independent steps, you’re adding complexity for no gain.

Examples that need agents:

Examples that don’t:

Red flag: If you can draw your process as a straight line with maybe a few branches, stick with automation.

3. Would a human need to make judgment calls?

The question: Are there places where someone with domain expertise would say “it depends” and need to consider factors that aren’t easily quantified?

Why it matters: Judgment calls are where agents can add real value, but they’re also where they can fail spectacularly. If your domain expertise can be encoded in rules, do that instead.

Examples that need agents:

Examples that don’t:

Red flag: If your “judgment calls” are really just complex rules you haven’t written down yet, write them down.

4. Do you need it to notice things and act on them?

The question: Should the system be watching for opportunities or problems and taking initiative, or is it fine to wait for human direction?

Why it matters: Proactivity is expensive to build and can go wrong in spectacular ways. Make sure the value of catching things early outweighs the risk of false alarms and misguided actions.

Examples that need agents:

Examples that don’t:

Red flag: If the cost of missing something is low, or if false positives would be more annoying than helpful, skip the proactivity.

5. How much does domain expertise actually matter?

The question: Would a smart generalist struggle with this task, or could they figure it out with access to the right information?

Why it matters: Deep domain knowledge is hard to encode and maintain. If your problem doesn’t really require specialized expertise, you might be overcomplicating things.

Examples where expertise matters:

Examples where it doesn’t:

Red flag: If you can train someone to do the task in a few weeks, the domain expertise requirement might not justify agent complexity.

Putting It Together: A Reality Check

Here’s how I actually use this framework: I go through the five questions and count how many get a strong “yes.”

4-5 strong yeses: You’ve probably got a genuine agent use case. The complexity is justified, and simpler approaches will likely fall short.

2-3 yeses: This is the gray area. You might want an agent, or you might want really good automation with some smart components. Consider starting simple and evolving.

0-1 yeses: You probably want automation, not an agent. The agent complexity will cost you more than it delivers.

But here’s the crucial part: even if you score high, that doesn’t mean you should build an agent. You also need to consider whether you have the team, the timeline, the budget, and the risk tolerance for what is inherently a more complex and unpredictable system.

Some Real Examples

Customer service for a SaaS company: Context changes (customer history, product issues), multiple reasoning steps (diagnosis, solution selection, escalation), requires judgment about tone and urgency, needs proactive follow-up, benefits from product expertise. Five yeses—probably worth exploring an agent approach, but start with human-in-the-loop.

E-commerce inventory management: Inventory levels and prices change constantly, requires sequential decisions about ordering and pricing, some judgment about seasonal patterns, benefits from proactive reordering, needs retail expertise. Four yeses—good agent candidate, but be careful about the proactive piece.

Employee onboarding: Mostly standard process, minimal context changes, straightforward step sequence, limited judgment required, reactive rather than proactive, minimal specialized knowledge needed. One yes—stick with a well-designed workflow tool.

Financial portfolio rebalancing: Market conditions change constantly, complex multi-step optimization, requires significant judgment about risk and timing, should be proactive about opportunities, deep financial expertise essential. Five yeses—but also high stakes, so probably want human oversight even with an agent.

When This Framework Fails

This thinking process works for operational workflows where you can define success reasonably clearly. It’s not great for:

In those cases, you’re probably better off with human-AI collaboration tools, specialized software, or just keeping humans in the loop.

The Real Question

The framework gives you a way to think through whether you need an agent, but the real question is often simpler: What’s the smallest thing you could build that would actually help?

Sometimes that’s a full autonomous agent. More often, it’s automation with a few smart components, or a human-in-the-loop system, or just better tooling for the people already doing the work.

Don’t let the excitement around agents push you toward complexity you don’t need. The best AI deployment is often the boring one that just works.


Edit page
Share this post on:

Previous Post
Book Signing Highlights: Decode and Conquer 5th Edition
Next Post
How to Get Decode and Conquer 5th Edition on Kindle, eBook, or PDF