<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Team-Structure on Damian Galarza | Software Engineering &amp; AI Consulting</title><link>https://www.damiangalarza.com/tags/team-structure/</link><description>Recent posts from Damian Galarza | Software Engineering &amp; AI Consulting</description><generator>Hugo</generator><language>en-us</language><managingEditor>Damian Galarza</managingEditor><atom:link href="https://www.damiangalarza.com/tags/team-structure/feed.xml" rel="self" type="application/rss+xml"/><item><title>Your AI Team Doesn't Need More People — It Needs Agents</title><link>https://www.damiangalarza.com/posts/2026-05-13-your-ai-team-doesnt-need-more-people-it-needs-agents/</link><pubDate>Wed, 13 May 2026 00:00:00 -0400</pubDate><author>Damian Galarza</author><guid>https://www.damiangalarza.com/posts/2026-05-13-your-ai-team-doesnt-need-more-people-it-needs-agents/</guid><description>What 'supported by a fleet of agents' means in practice: which tasks automate, which don't, and where the ROI breaks down. Evidence from Stripe, Coinbase, Ramp, and Shopify.</description><content:encoded><![CDATA[<p>Stripe merges 1,300 agent-written PRs per week. Every one is human-reviewed. None contain human-written code.</p>
<p>That number showed up in <a href="https://stripe.dev/blog/minions-stripes-one-shot-end-to-end-coding-agents-part-2">Stripe&rsquo;s engineering blog</a> in March 2026, and it has been circulating in boardroom conversations ever since. Not because it&rsquo;s impressive as a volume metric, but because of what it implies about the relationship between team size and engineering output. If a fleet of agents can produce that volume of shippable work, and humans are reviewing rather than writing, the old model of scaling output by scaling headcount starts to look like the wrong optimization.</p>
<p>This is already reshaping how engineering leaders think about team structure. Leigh Newsome at <a href="https://hoolahoop.io/articles/cto-coaching/ai-native-team-topology/">HoolaHoop</a> captures the pattern emerging across mid-to-late-stage companies: three to five senior engineers, supported by agents, matching or exceeding what eight to twelve person teams shipped a year ago. It&rsquo;s showing up in the organizations they coach, not just the headline reference companies.</p>
<p>Coinbase is taking it further. Brian Armstrong&rsquo;s <a href="https://x.com/brian_armstrong/status/2051616759145185723">May 2026 restructuring announcement</a> — cutting fourteen percent of the workforce amid a crypto downturn — reorganizes the company around what he calls &ldquo;AI-native pods.&rdquo; Small teams where engineers, designers, and product managers collapse into single roles, managing fleets of agents. No pure managers. Five layers max below CEO/COO. &ldquo;One person teams&rdquo; as an explicit experiment. The framing is blunt: &ldquo;rebuilding Coinbase as an intelligence, with humans around the edge aligning it.&rdquo; You can debate whether the layoff is cost-cutting dressed as innovation or genuine structural change. But the organizational thesis — that small, high-context teams with agent leverage replace larger ones — is the same pattern showing up everywhere else.</p>
<p>The question most engineering leaders are asking right now isn&rsquo;t whether this shift is real. It&rsquo;s what &ldquo;supported by a fleet of agents&rdquo; actually means in practice.</p>
<h2 id="why-the-old-math-broke">Why the old math broke</h2>
<p>For twenty years, engineering scale was synonymous with headcount. Need more output? Hire more engineers. Want to ship faster? Add a team. The org chart grew, the roadmap grew, and the relationship between the two felt roughly linear.</p>
<p>That linear relationship depended on an assumption: execution required human hands. Writing code, running tests, fixing lint errors, updating documentation, opening PRs. Each of these was a unit of labor that only a person could perform. So more labor required more people.</p>
<p>Agents broke that assumption. Not for all tasks, but for enough of them to change the math.</p>
<p>The shift isn&rsquo;t about replacing engineers. It&rsquo;s about changing what makes an engineering team productive. A senior engineer who can spec work clearly, verify agent output efficiently, and maintain the infrastructure that agents run on produces more than a senior engineer writing all the code themselves. Add two more seniors with the same skills, plus a fleet of agents executing the predictable work, and you have a small team with the output of a much larger one.</p>
<p>This is a unit-economics revision, not a layoff strategy. The teams getting this right are redirecting budget from planned mid-level hires toward senior talent and agent infrastructure. The ones getting it wrong are either cutting headcount and hoping agents fill the gap (they won&rsquo;t) or ignoring the shift entirely and wondering why smaller teams elsewhere are shipping faster.</p>
<h2 id="what-plus-agents-means-in-practice">What &ldquo;plus agents&rdquo; means in practice</h2>
<p>The phrase &ldquo;supported by agents&rdquo; is easy to say and hard to make concrete. Here&rsquo;s how I think about it: some tasks automate well with current agent capabilities, and some don&rsquo;t. Knowing the difference is the entire game.</p>
<h3 id="tasks-that-automate-well">Tasks that automate well</h3>
<p><strong>Boilerplate and scaffolding.</strong> Generating new files, wiring up routes, creating database migrations, adding API endpoints that follow existing patterns. Agents handle this reliably because the work is repetitive and the success criteria is structural, not judgmental.</p>
<p><strong>Test generation.</strong> Writing unit tests for existing code, especially when the code follows clear patterns. Agents are surprisingly good at identifying edge cases when given a well-defined function signature and existing test examples to follow.</p>
<p><strong>Refactoring.</strong> Renaming across a codebase, extracting modules, updating import paths, migrating from one API version to another. Deterministic transformations with verifiable outcomes.</p>
<p><strong>Lint, formatting, and CI fixes.</strong> This is where Stripe&rsquo;s Blueprint pattern shines. They run lint as a deterministic node in their agent state machine before pushing to CI. The agent doesn&rsquo;t need judgment here, just compliance.</p>
<p><strong>Documentation updates.</strong> Syncing docs with code changes, updating README files, regenerating API references. The source of truth (the code) already exists; the agent just needs to reflect it.</p>
<h3 id="tasks-that-dont-automate-well">Tasks that don&rsquo;t automate well</h3>
<p><strong>Architecture decisions.</strong> Which database to use, how to structure service boundaries, when to split a monolith. These require context that lives in organizational history, business constraints, and operational experience. No amount of context window will replace knowing what broke in production last quarter.</p>
<p><strong>Product judgment.</strong> Deciding what to build, what to cut, what to defer. Agents can execute against a spec, but they can&rsquo;t decide whether the spec solves the right problem.</p>
<p><strong>Spec authorship.</strong> Writing the acceptance criteria that agents execute against. This is the inverse of what you might expect. Vague specs produce vague work, fast. The cost of ambiguity in product input is amplified by agents, not absorbed by mid-level engineers the way it used to be. Spec quality is now directly proportional to output quality.</p>
<p><strong>Verification.</strong> Confirming that agent-generated code actually solves the problem it was supposed to solve. This is the new bottleneck, and it deserves its own section.</p>
<h2 id="where-the-roi-breaks-down">Where the ROI breaks down</h2>
<p>Agent leverage is real, but it compounds unevenly. Three areas determine whether you&rsquo;re getting actual value or just generating volume.</p>
<h3 id="the-verification-bottleneck">The verification bottleneck</h3>
<p>Agent output is fast but inconsistent. A fleet of agents can generate fifty PRs in an afternoon. If each one takes a senior engineer thirty minutes to verify, you&rsquo;ve consumed twenty-five hours of your most expensive engineering time reviewing code instead of designing systems.</p>
<p>This is where new metrics start to matter. <a href="https://optimumpartners.com/insight/engineering-management-2026-how-to-structure-an-ai-native-team/">Optimum Partners</a> proposes three that I think are worth tracking:</p>
<p><strong>MTTV (Mean Time to Verification)</strong> measures how long it takes to confirm that agent output is actually correct and ready to ship. This replaces &ldquo;code complete&rdquo; as the meaningful quality milestone. If MTTV is high, you&rsquo;re not getting leverage even when volume looks impressive.</p>
<p><strong>AI-CFR (AI Change Failure Rate)</strong> tracks the percentage of agent-generated changes that fail in production or require post-merge rework. This needs to be tracked separately from human change failure rate because the failure modes are different. Agent-generated code tends to fail in different ways than human-written code, and combining the metrics hides the problem.</p>
<p><strong>Interaction Churn</strong> measures how many back-and-forths a human has with an agent before reaching a useful result. High churn signals weak specs, poor context engineering, or the wrong agent for the task. It&rsquo;s the leading indicator of whether your AI investment is actually compounding.</p>
<p>The uncomfortable implication: if your engineering review next quarter uses none of these metrics, you&rsquo;re running an AI-curious team, not an AI-native one.</p>
<h3 id="infrastructure-quality-compounding">Infrastructure quality compounding</h3>
<p>Here&rsquo;s something that gets missed in most conversations about agent teams. Stripe didn&rsquo;t become good at agents overnight. They spent years investing in developer experience. Their development environments spin up in ten seconds. Their MCP Toolshed contains <a href="https://stripe.dev/blog/minions-stripes-one-shot-end-to-end-coding-agents-part-2">nearly 500 tools</a> in a centralized registry. Their rule files are scoped to subdirectories rather than global, because global rules in large repos waste the agent&rsquo;s context window before the agent starts working.</p>
<p>The insight from their engineering team is deceptively simple: <a href="https://stripe.dev/blog/minions-stripes-one-shot-end-to-end-coding-agents-part-2">&ldquo;What&rsquo;s good for humans is good for agents.&rdquo;</a> Every investment they made in human developer productivity — fast feedback loops, clean tooling, isolated environments — returned dividends when agents started running the same tooling.</p>
<p>This cuts both ways. Teams that cut corners on developer experience are discovering those shortcuts at ten times the speed when agents surface them. If your codebase is messy, agents will reproduce that mess faster than any human could. If your CI pipeline is flaky, agents will burn cycles retrying it. If your test coverage is thin, agent-generated code will ship without guardrails.</p>
<p>Infrastructure quality isn&rsquo;t just a nice-to-have in the agent era. It&rsquo;s a force multiplier. The gap between well-maintained and poorly-maintained codebases is widening faster than ever.</p>
<h3 id="spec-precision">Spec precision</h3>
<p>Ramp&rsquo;s story adds a different dimension. Their AI usage grew 6,300% year-over-year. 99.5% of their team is active on AI tools. 84% use coding agents weekly. And here&rsquo;s the number that matters most: non-engineers now account for <a href="https://x.com/geoffintech/status/2042002590758572377">12% of all human-initiated PRs on Ramp&rsquo;s production codebase</a>, thousands per month.</p>
<p>They got the org design wrong before they got it right. The initial instinct was to centralize: one small team builds tools for the whole company. Demand outstripped capacity immediately. Then they swung decentralized. Every team builds its own solutions. Redundant re-learning everywhere.</p>
<p>The answer was hub-and-spoke. A small central team builds the platforms, connectors, and infrastructure. Functional teams build solutions on top and feed requirements back. A risk analyst automated sixteen hours per month of manual modeling. A sales ops lead replaced a spreadsheet-based comp model across three orgs in forty-eight hours. An L&amp;D lead built a training simulator in fifteen minutes. None of them were engineers.</p>
<p>This works because the central team invested in spec precision. Their internal AI workspace, <a href="https://x.com/eglyman/status/2043362828178841860">Glass</a>, auto-configures on install, connecting thirty-plus tools through single sign-on. No setup guide. No IT ticket. When the barrier between &ldquo;I have a problem&rdquo; and &ldquo;I can spec a solution&rdquo; drops to near zero, the bottleneck shifts from execution to specification quality.</p>
<p>For agent fleets specifically, the same principle applies. Centralize the infrastructure, decentralize the application. And invest relentlessly in making specs clear enough that agents can execute against them without ambiguity.</p>
<h2 id="juniors-arent-going-away">Juniors aren&rsquo;t going away</h2>
<p>The surface-level reading of this story says junior engineers are obsolete. The data contradicts it.</p>
<p>Shopify is publicly expanding its junior pipeline, going from <a href="https://coderpad.io/blog/hiring-developers/in-the-ai-era-shopify-is-investing-in-junior-engineers-not-cutting-them/">roughly a hundred interns to over a thousand a year</a>, while reporting around a <a href="https://www.bvp.com/atlas/inside-shopifys-ai-first-engineering-playbook">twenty percent productivity lift</a> from AI tooling. That&rsquo;s not a company that thinks juniors are dead weight.</p>
<p>What&rsquo;s happening is subtler. Junior engineering work is being repointed. Instead of writing boilerplate code that agents now handle, juniors are being directed toward verification, spec authoring, and the kind of detail-oriented review work that AI-native teams need more of. The AI Reliability Engineer role that&rsquo;s emerging in some organizations is essentially a formalization of this shift: someone who validates agent-generated output and designs the verification systems that make AI work safely shippable.</p>
<p>Entry-level engineering isn&rsquo;t disappearing. The definition of entry-level work is changing. The junior who can verify agent output, write precise acceptance criteria, and identify when an agent&rsquo;s solution is structurally correct but semantically wrong is exactly the junior AI-native teams need.</p>
<h2 id="where-to-start">Where to start</h2>
<p>If you&rsquo;re leading an engineering team and this resonates, here&rsquo;s what I&rsquo;d do first.</p>
<p><strong>Audit your developer experience.</strong> Before adding agents to your workflow, look at what they&rsquo;ll inherit. Is your CI fast? Are your test environments isolated? Is your documentation current? Agents amplify whatever they find. Make sure what they find is solid.</p>
<p><strong>Measure verification, not velocity.</strong> Track the three metrics from the verification section: MTTV, AI-CFR, and Interaction Churn. These tell you whether you&rsquo;re getting leverage or just generating volume.</p>
<p><strong>Automate the predictable, keep humans on the judgment.</strong> Map your team&rsquo;s work to the two lists above. If a task is repetitive, structurally verifiable, and follows existing patterns, it&rsquo;s a candidate for agent automation. If it requires organizational context, product judgment, or architectural reasoning, it stays with humans.</p>
<p><strong>Invest in spec quality.</strong> The biggest single improvement most teams can make is writing clearer specs. When agents execute against precise acceptance criteria, the output improves dramatically. When they execute against vague user stories, you get technically-shipped work that doesn&rsquo;t solve the problem.</p>
<p><strong>Stop hiring against pre-agent plans.</strong> If the headcount plan you presented to the board was built before agents changed the unit economics, revisit it. The next ten engineers you would have hired might be better allocated as three senior engineers and an investment in agent infrastructure. Frame it as a unit-economics revision, not a hiring miss.</p>
<p>None of this is theoretical. The companies producing evidence (Stripe&rsquo;s 1,300 PRs per week, Coinbase restructuring around AI-native pods, Ramp&rsquo;s 12% non-engineer PRs, Shopify&rsquo;s expanded junior pipeline) are operating at a different leverage ratio than teams still sizing by headcount alone. The gap is measurable and it&rsquo;s growing.</p>
<h2 id="working-with-me">Working with me</h2>
<p>I help founders and engineering teams navigate the shift where the old headcount math stops making sense. We figure out where agents fit, where humans still need to own judgment, and what your team has to change to ship reliably. If that conversation would be useful, <a href="/services/ai-engineering/">let&rsquo;s talk</a>.</p>
<h2 id="further-reading">Further reading</h2>
<ul>
<li><a href="https://stripe.dev/blog/minions-stripes-one-shot-end-to-end-coding-agents-part-2">Minions: Stripe&rsquo;s one-shot, end-to-end coding agents, Part 2</a> (Stripe Dev Blog)</li>
<li><a href="https://x.com/geoffintech/status/2042002590758572377">How to get your company AI pilled</a> (Geoff Charles, Ramp)</li>
<li><a href="https://hoolahoop.io/articles/cto-coaching/ai-native-team-topology/">AI-Native Team Topology</a> (HoolaHoop, Leigh Newsome)</li>
<li><a href="https://x.com/brian_armstrong/status/2051616759145185723">Coinbase restructuring memo</a> (Brian Armstrong)</li>
<li><a href="https://x.com/eglyman/status/2043362828178841860">Glass announcement</a> (Eric Glyman, Ramp)</li>
<li><a href="https://www.bvp.com/atlas/inside-shopifys-ai-first-engineering-playbook">Inside Shopify&rsquo;s AI-First Engineering Playbook</a> (Bessemer Venture Partners)</li>
</ul>
]]></content:encoded></item></channel></rss>