Background
As with most engineering organizations, "Adopting AI" became a mandate for our company sometime this year. No one knew what it really meant to be AI first but one thing was clear: Leadership expected engineering teams to lead the way on this initiative. LinkedIn is rife with claims of engineers single handedly shipping entire products in weeks, sometimes days or even while on a Zoom call. One has to pay attention when PG starts posting this, right? Teams then move at light speed to procure licenses for whatever tools their leadership, board, peers, HN/Reddit posts and "experts" recommended.
And then comes the hard (and fun) part.
If you're like most engineering teams, then you saw a pattern. After the initial euphoria settled down and the teams got some ostensible wins, they didn't move at the amazing speed leaders expected. Bug count, regressions, PR review times, and release velocity barely moved despite every engineer burning hundreds of dollars worth of tokens daily.
This happens because teams need more than guidance; they need goals. After observing multiple engineering teams navigate AI adoption, I observed a clear pattern: successful teams progress through three distinct phases in their AI adoption journey — Minimalist, Collaborative, and Accelerating. Each phase has specific goals, metrics to track those goals, investment requirements, and transition triggers.
I'll introduce the three phases in this post and explain some behavioral traits of being in each phase
Defining the problem
Before I dive into the phases, I'd like to share my thesis on why, despite immense adoption, around 95% of generative AI pilots are failing.
I view all new AI initiatives as split into two broad categories: Internal and Product.
Internal: These initiatives aim to improve the overall output of the organization with same headcount. "Do more with less" seems to be the prevailing motto here. This memo by Shopify CEO captures the sentiment pretty well.
Product: These initiatives aim to generate more revenue by increasing the product inventory, simplifying integrations, improving existing features, improve usability, expanding into new domains or all of the above.
Although the framework I'm working on focuses on Internal initiatives, and more specifically, for engineering teams and their stakeholders, it can be applied to the Product initiatives as well. I'll share the efficacy of this framework on Product initiatives once my experiments on that front reach their conclusion.
There are two major problems that teams are facing with gen AI adoption, assuming teams have the leadership buy-in to do this.
1. Non-fungible learnings, success stories, runbooks, guidance and processes. Successful AI projects include both technical adoption as well as team, process, culture specific integrations. These teams codify their communication patterns, knowledge silos, hacks, runbooks and knowledge bases to build unique solutions which cannot be adopted as-is for anyone else.
2. Lack of a cohesive approach. Any AI adoption framework must ensure that the individual productivity gains compound into team effectiveness, ultimately driving company-wide transformation. Efficiency gains only compound when they're aligned across individuals and teams. Lack of a converging approach results in productivity silos.
Consider two scenarios I've encountered repeatedly:
Scenario 1: An 8-person team using Cursor and Claude Code sees initial excitement but no sustained gains. Their metrics—bugs, regressions, velocity—remain unchanged after three months. Engineers report lifechanging productivity gains but key metrics remain stagnant.
Scenario 2: A mid-size organization with multiple teams struggles with database consistency. One team liberally uses database transactions while another uses them sparingly—for nearly identical microservices. Their code can't effectively merge or integrate. This creates operational overhead, capacity planning inconsistencies and architecture silos.
In both scenarios, knowledge silos prevent measurable, impactful organizational improvements. Using AI tools without alignment just speeds up the pace at which teams head towards a conflict. A unifying framework helps individuals, teams and organizations bridge this gap. This doesn't mean prescribing the "right" way to adopt AI. It means providing them a framework for self-discovery while maintaining alignment.
AI Adoption Framework
The framework I built to successfully adopt AI across engineering teams builds on three phases, each focusing on the "What to do" rather than "How to do it". Each phase in this framework builds on top of the previous phase to compound the learnings into exponential gains.
I created this AI Adoption Framework through multiple iterations and experiments. I'll continue to share my learnings as we explore new methods that work and the ones that don't.
Phase 1, Minimalist
This phase is indicated by siloed AI adoption across individuals, processes and teams.
Teams begin by augmenting existing tools with AI capabilities. Engineers replace traditional IDEs with Cursor or Windsurf. They adopt Claude Code to orchestrate changes across multiple files simultaneously. Engineers enhance code reviews through Copilot, Bugbot or similar tools. Engineers start creating custom prompts and GPTs to solve team-specific or product-specific problems.
At this stage, engineers evolve from coders to orchestrators. Engineers no longer type every line—they direct AI tools to implement their vision. Most engineering teams are already here or getting here fast.
Measurable Checkpoints:
- 80% of engineers save 10-20% time daily
- Sprint velocity increases by measurable story points
- Lines of code per engineer increases
- Number of tickets closed per sprint trends upward
Phase 2, Collaborative
Defined by collaboration between teams and business units using AI tools and with AI tools.
Teams now transform individual productivity gains into team-wide efficiency through codified expertise. This phase represents the first step toward agentic workflows. Knowledge sharing becomes embedded in prompts, CLAUDE.md files, and shared artifacts.
Consider John, a DevOps expert who spends 60% of his day reviewing infrastructure PRs. John creates a Claude Code command: /john-infra-review
and commits it to the project's .claude/commands
folder.
The command contains:
- All infrastructure review guidelines John follows and enforces
- Patterns extracted from analyzing five months of his PR comments
- Focus areas identified through LLM analysis of his review history
- Specific checks for security, scalability, and cost optimization
Now any team member runs this command before raising an infrastructure PR. They receive feedback as if John personally reviewed their code. John's review load drops by 70%, PRs arrive pre-vetted, and, he focuses on important architectural decisions rather than routine checks and formatting errors.
This slash command creates the foundation for John's future autonomous agent.
Measurable Checkpoints:
- PR review time reduces by 40%. Indicates increased collaboration before sending the PR
- Average and 90th percentile review cycles count drop
- Cross-team collaborative projects complete 25% faster
- Delivery time for features involving multiple engineers decreases
Phase 3, Accelerated
Agentic and autonomous systems collaborating to drive business outcomes
Teams now create autonomous agents that deliver non-trivial outcomes with the same headcount. The goal: make your team execute at 1.5x-2.0x its previous capacity.
Consider an engineer writes a detailed Jira ticket with clear outcomes. Their personal agent:
- Picks up the task automatically
- Analyzes requirements and creates implementation plan
- Makes necessary codebase changes
- Submits changes to review agents created by teammate(s) in Phase 2
- Processes feedback and iterates on the solution
- Re-submits updated changes for agent review
- Only after passing all automated reviews, the agent creates a PR and tags human reviewers
Engineers evolve into master orchestrators, focusing on architecture, design, cost impact, business alignment, and code quality standards. Teams automate mundane implementation details — the "how" — while humans focus on the "what" and "why"
Measurable Checkpoints:
- 30% of tasks complete without human intervention on first iterations
- Regression rate drops by 45%
- Deployment frequency doubles
- Team delivers 1.5x-2.0x previous output, measured MoM
- Percentage of tasks requiring human input decreases from 100% baseline
Conclusion
I've briefly introduced the AI adoption framework I'm using and experimenting with in this post.
In the next series of posts, I'll dive into mechanisms I used with my teams to successfully move through each stage. I'll also talk about common symptoms of teams that are stuck in a certain stage and how to move past those blockers.
Stay tuned!