Treating an AI like an employee: what actually changes
The framing has become a cliché. The actual mechanics, onboarding, scope of work, performance review, termination, change in interesting ways once you take the metaphor literally.
"Treat your AI agent like an employee" is one of those framings that sounds insightful the first time and tired the fifth time. It usually shows up in the same essay as "AI doesn't replace people, it makes them more productive" and ends with a paragraph about prompt engineering being the new literacy. The framing isn't wrong. It's mostly just lazy in how it gets used.
The version of the framing that actually pays off is the one that takes the metaphor literally enough to extract the operational mechanics from it. Onboarding. Scope of work. Performance review. Termination. Each of those moves from being a cliché bullet point to being a concrete change in how the work happens, once you do it deliberately. Worth being specific about each.
Onboarding
When a new engineer starts, you don't hand them root access on day one and tell them to figure it out. You give them a curated set of permissions, a project that's bounded enough to learn from, written context about how the team works, and a senior person who reviews everything they touch for the first few weeks. The senior person isn't there because the new engineer is incompetent; they're there because trust has to be earned at the boundary where the new engineer's work meets shared infrastructure.
The same shape applies to an AI agent in a new project, and almost nobody does it. The default workflow is "install the agent, give it your repo, let it see everything, and start asking it to do work." That's the AI-equivalent of giving the new engineer root and walking away.
A useful onboarding pattern that's emerged in the first month of working with an agent in your editor:
- Start the agent in a worktree or branch the agent can't merge from. The agent works in isolation; the human reviews and merges.
- Give the agent a context document, a CLAUDE.md or AGENTS.md or whatever the conventions for this tool are, that says explicitly what this codebase cares about, what conventions matter, and what's off-limits.
- Give it one bounded task to start with, a refactor, a small feature, a documentation pass. Not "improve this codebase."
- Review the output line-by-line for the first few sessions. Same way you'd review a new engineer's first PRs.
The investment is real and pays off the same way: the agent gets calibrated to your context, you get calibrated to its failure modes, and the relationship stops being "type things and hope for the best."
Scope of work
A new employee doesn't show up with an undefined job description. They have a role, a manager, and a set of responsibilities. The boundaries get adjusted as trust builds, but they exist from the start.
The AI-as-employee equivalent is being explicit about what the agent's role is on a project. Is it the code generator and the human is the architect? Is it the reviewer and the human is the author? Is it the documenter and the human writes the actual code? These aren't interchangeable, and trying to make the agent do all three at once produces worse results than picking one and being deliberate about it.
The most concrete useful change once you make the role explicit: prompt patterns become reusable. "You are the code reviewer for this project. Your job is X, Y, Z. You don't do A, B, C." That kind of system prompt, scoped to a role, is a fundamentally different artifact from "please help me with my code." It's something you save, version, and reuse. The agent without a role is a friendly stranger; the agent with a role is something closer to a colleague with a job description.
Performance review
Real performance management isn't just "is this person good or bad". It's a recurring look at what's working, what isn't, and what to adjust. The cadence and the specifics matter.
For an AI agent in a project, the equivalent is regular review of the agent's actual output and adjustment of the prompt patterns and tool configurations that produced it. A pattern I've found useful: at the end of every couple of weeks of agent-assisted work, look at the commits the agent contributed to, sort them by how much rework they needed, and ask which classes of task the agent kept getting wrong. The answer for a coding agent is usually consistent, there are categories where it's reliably good and categories where it's reliably bad, and the failure modes are stable.
That insight then feeds the next iteration. Maybe the project's CLAUDE.md needs more detail about the test patterns. Maybe certain kinds of refactors should be done by hand. Maybe a tool the agent has access to needs better documentation in its name and description so the agent picks it correctly. The performance review is the mechanism that makes those adjustments deliberate rather than ad hoc.
Termination
Knowing when to fire a tool is underrated. With a human employee, the cost of keeping a bad fit around includes the work they don't do well, the morale impact on the team, and the opportunity cost of the better hire you could be onboarding instead. With an AI tool, the cost is more abstract but real: every workflow that depends on the tool is harder to migrate the longer you've depended on it; every habit that formed around the tool's quirks is a habit you'll have to unlearn.
The version of "termination" that comes up most in practice is not getting rid of the category of agent (those are mostly here to stay) but knowing when to switch from one specific tool to another. When a new product like Claude Code lands and changes the shape of what's possible, the right move is sometimes to migrate even if the previous tool was working. The wrong move is to migrate every time something new comes out, because the cost of switching includes re-onboarding effort that took weeks to land in the first place.
A working heuristic: switch tools when the new one is meaningfully better at the category of work you're doing, not when it's marginally better at a benchmark. The benchmark improvements compound for vendors; the category-of-work improvements are the ones you actually feel.
What the metaphor doesn't extend to
The framing breaks down at a few specific places, and being honest about them matters:
- An AI agent doesn't have a career arc. A junior engineer joins, learns, gets promoted, eventually moves on. None of that maps. The agent is what it is on the day you install it, and the model behind it changes when the vendor decides, not when you decide.
- There's no peer relationship. An employee learns from teammates and contributes to their growth. The agent does neither. Whatever knowledge transfer happens is one-way and contingent on the human bothering to extract it.
- The agent doesn't have skin in the game. A bad outcome is your problem. The agent moves on to the next prompt, no awkward conversation, no blocked promotion, no team dynamics impact. That asymmetry is why the trust calibration has to be done so deliberately, there isn't an analogous self-correction mechanism.
- The metaphor encourages over-trust. The most consistent failure mode I've seen, my own included, is starting to treat the agent as if it has the kind of judgment a teammate would, and being surprised when it doesn't. Useful framing in moderation, dangerous framing taken too literally.
The honest version of "treat the AI like an employee" is more like treat the AI like a contractor who's smart and fast but who doesn't actually understand the business and won't tell you when they're guessing. That's a different management mode than treating it like a team member, and the specifics (the scope of work, the review cadence, the willingness to fire it from a task) follow from there.