1Week 1-2 · ~4 hours

Beyond Scrum Basics

Scrum is a fine starting point, but most mature teams outgrow it. This module covers what comes next: Scrumban, flow-based delivery, probabilistic forecasting, and a Definition of Done that actually means something.

When Pure Scrum Breaks Down

Let me be direct: Scrum works well for roughly 18 months. A new team adopts it, gets the ceremonies going, starts delivering predictably, and leadership is happy. Then something shifts. The sprints start feeling like artificial containers. The team is spending more time in ceremonies than building software. Stories keep bleeding across sprint boundaries. Velocity becomes a weapon rather than a diagnostic tool.

I have seen this pattern play out across dozens of teams. The symptoms are predictable: sprint planning takes 3+ hours because the backlog is a mess, standup becomes a status report to the Scrum Master rather than a team sync, and the retrospective is the same three complaints on repeat. The team is doing Scrum without being agile.

The root cause is usually one of three things. First, the work does not fit neatly into 2-week boxes. Support teams, platform teams, and ops-heavy teams deal with a constant stream of unplanned work that makes sprint commitments meaningless. Second, the team has matured past the need for prescribed ceremonies. They coordinate naturally and the overhead of formal events creates friction rather than alignment. Third, the environment demands faster feedback loops than a 2-week sprint provides. When you are deploying 5 times a day, the concept of a “sprint review” at the end of a 2-week cycle feels absurd.

This is not a failure. It is growth. The mistake is clinging to pure Scrum when the team has outgrown it. Scrum is training wheels, and training wheels are great until they slow you down.

Scrumban: The Practical Middle Ground

Scrumban is not a framework you install. It is an evolution you grow into. The core idea is simple: keep the parts of Scrum that serve you (sprint cadence for planning, retrospectives for improvement) and layer in Kanban practices (WIP limits, pull-based flow, explicit policies) to address the parts that do not.

In practice, a Scrumban transition looks like this. You keep your sprint cadence but stop making sprint commitments. Instead, you use the sprint boundary as a planning trigger: when the sprint ends, the team replenishes the board with enough work for the next cycle. There is no “sprint goal” in the traditional sense. There is a prioritized queue and a pull-based system.

You add WIP limits to your board columns. This is the single most impactful change you can make. A typical starting point:

Recommended Starting WIP Limits

In Progress (Dev)Team size minus 1 (e.g., 5 devs = WIP 4)
Code ReviewTeam size divided by 2 (e.g., 5 devs = WIP 2-3)
QA / Testing2-3 items max regardless of team size
Awaiting Deploy3-5 items (triggers a deployment if exceeded)

The magic of WIP limits is counterintuitive: by doing less at once, you finish more. This is not philosophy; it is queuing theory. When a developer is juggling 4 items, context-switching costs eat 20-40% of their productive time. When they focus on 1 item, that cost drops to near zero.

Standup changes too. Instead of each person reporting what they did yesterday, the team walks the board right to left. “What is closest to done? What is blocked? Where are we exceeding WIP limits?” This reframes the conversation from individual status to system flow. It typically takes 5-8 minutes instead of 15-20.

The Science Behind WIP Limits: Little's Law

WIP limits are not a gut-feel heuristic. They are grounded in a mathematical principle called Little's Law, formulated by John Little in 1961. The formula is deceptively simple:

Lead Time = WIP ÷ Throughput

Or equivalently: WIP = Throughput × Lead Time

In plain English: if your team finishes 10 items per week (throughput) and you have 20 items in progress (WIP), your average lead time is 2 weeks. Want to cut lead time in half? You have two options: double your throughput (hard, expensive, usually impossible) or halve your WIP (free, immediate, consistently effective).

This is why adding more work to a busy team makes them slower, not faster. Every additional item in the system increases lead time for every other item. I once worked with a team that had 47 items “in progress.” Their throughput was about 8 items per week, meaning average lead time was nearly 6 weeks. We cut WIP to 12. Throughput stayed at 8 (it often even increases slightly because of reduced context-switching). Lead time dropped to 1.5 weeks. Same team, same people, same capacity. Just fewer plates spinning.

The practical implication for PMs: stop managing scope through sprint commitments and start managing flow through WIP limits. When a stakeholder asks “when will this be done?,” you can give a probabilistic answer based on your current WIP and throughput rather than a fiction based on story point estimates.

Why Story Points Are Broken (And What to Use Instead)

Story points were invented by Ron Jeffries as a way to decouple estimation from time, letting teams express relative complexity without committing to hours. The idea was sound. The execution, across the industry, has been a disaster.

Here is what goes wrong. Management starts treating story points as a productivity metric. “Team A delivers 40 points per sprint and Team B delivers 25. Why is Team B slower?” This question is meaningless because points are relative to each team, but it gets asked in every leadership meeting. Teams respond rationally by inflating their estimates. A task that was a 3 becomes a 5. Velocity goes up. Everyone is happy. Nothing actually changed.

The deeper problem is that estimation accuracy does not improve with practice. Research by Vasco Duarte and others has shown that most teams cannot estimate more accurately than “small, medium, or large.” The difference between a 5 and an 8 is noise. Yet teams spend hours in planning poker debating whether something is a 5 or an 8, time that could be spent actually building the software.

The Alternative: Cycle Time Forecasting

Instead of estimating how big something is, measure how long things actually take. Cycle time is the clock time from when work starts to when it is done. Collect this data for 8-12 weeks and you have a distribution that is far more predictive than any estimate.

A typical team might find: 50% of their items complete in 3 days or less, 85% complete in 7 days or less, and 95% complete in 14 days or less. That 95th percentile number is your “safety margin.” When a stakeholder asks for a delivery date, you can say: “Based on our historical data, there is an 85% chance this will be done within 7 working days.”

Monte Carlo Simulations for Multi-Item Forecasting

For larger scopes (a feature with 15 items, a quarterly plan with 80 items), Monte Carlo simulation is remarkably powerful and surprisingly easy. The process works like this:

  1. Collect your throughput data: how many items did the team complete each week for the last 12+ weeks?
  2. Randomly sample from that historical data to simulate future weeks. If the team completed 6 items one week, 9 another, and 4 another, randomly draw from those actuals.
  3. Run the simulation 10,000 times. Each run gives you a projected completion date.
  4. The distribution of completion dates gives you confidence intervals: “80% of simulations finish by March 15, 95% finish by March 28.”

You can build this in a spreadsheet in 30 minutes. Tools like ActionableAgile and Nave automate it. The output is a probabilistic forecast grounded in actual performance data, not wishful thinking. I have used this technique to forecast delivery dates for programs with 200+ items across 6 teams and consistently landed within the 85% confidence interval.

The Four Levels of Definition of Done Maturity

The Definition of Done is the single most underrated artifact in Agile. Most teams treat it as a checkbox exercise during team formation and never revisit it. But your DoD is effectively your quality contract with the organization. It defines the minimum bar for what “done” means, and raising that bar is one of the highest-leverage improvements a team can make.

I categorize DoD maturity into four levels. Most teams operate at Level 1 or 2 and think they are at Level 3.

Level 1

“It Compiles”

Code is written, builds pass, basic happy-path testing is done by the developer. PR is merged. This is where most junior teams start. The problem: bugs are found in QA or production, rework is high (typically 30-40% of sprint capacity), and “done” items keep coming back. You will recognize this level by the phrase “it worked on my machine.”

Level 2

“It's Tested”

Code has unit tests, integration tests pass, code review is completed, and QA has verified the acceptance criteria. This is the standard most teams aspire to. The gap: testing is still a separate phase, non-functional requirements (performance, accessibility, security) are afterthoughts, and documentation is “we will do it later” (which means never).

Level 3

“It's Releasable”

Everything in Level 2, plus: feature is behind a feature flag and tested in a staging environment, documentation is updated (API docs, runbooks, user-facing help), performance testing confirms no regressions, accessibility standards are met, and security review is complete for sensitive changes. The key distinction: any item that meets this DoD can go to production right now without additional work.

Level 4

“It's Monitored in Production”

Everything in Level 3, plus: the feature has observability built in (structured logging, metrics, alerts), dashboards are updated to track the new functionality, rollback procedures are documented and tested, and the team has confirmed the feature works correctly in production with real user traffic. “Done” means “validated in production,” not just “deployed to production.”

Moving from Level 1 to Level 2 typically takes 2-3 months. Level 2 to Level 3 takes 3-6 months and requires investment in CI/CD, feature flags, and automated testing infrastructure. Level 3 to Level 4 requires a cultural shift where the team owns their code in production, not just in development.

The payoff is massive. Teams operating at Level 4 have near-zero production incidents caused by new deployments, because every release is small, tested, monitored, and reversible. They deploy multiple times per day with confidence. Their rework rate drops below 10%. This is what “engineering excellence” actually looks like in practice.

Flow Metrics That Actually Predict Delivery

Velocity is a lagging indicator. It tells you what happened, not what will happen. And because it is based on estimates (story points), it is built on a foundation of imprecision. Here are the four flow metrics that actually predict whether you will deliver on time:

1. Cycle Time (Lead Time for Changes)

The elapsed time from work starting to work completing. Track the 50th, 85th, and 95th percentiles. If your 85th percentile cycle time is increasing, you have a systemic problem regardless of what velocity says.

2. Throughput

The number of items completed per unit of time (typically per week). This is the input to Monte Carlo simulations and the most honest measure of team capacity. Unlike velocity, it is not gameable.

3. Work Item Age

How long has each in-progress item been in progress? This is a leading indicator. If an item has been in progress for longer than your 85th percentile cycle time, it is at risk and needs attention now, not at the next standup.

4. Work In Progress (WIP)

The count of items currently in progress. Per Little's Law, this directly determines your lead time. If WIP is trending up and throughput is flat, you are heading for trouble.

Together, these four metrics form a complete picture of your delivery system. I build a weekly dashboard with these numbers for every team I work with. It takes 10 minutes to update and replaces the 2-hour “are we on track?” meetings that plague most programs.

Key Takeaways

  • Pure Scrum is a starting point, not a destination. When it starts creating friction rather than reducing it, evolve toward Scrumban with WIP limits and pull-based flow.
  • Little's Law is not optional. WIP divided by throughput equals lead time. Reducing WIP is the fastest, cheapest way to reduce delivery time.
  • Stop debating story points. Use cycle time data and Monte Carlo simulations to produce probabilistic forecasts grounded in reality, not estimates grounded in hope.
  • Your Definition of Done is your quality contract. Invest in raising it from “it compiles” to “it's monitored in production.” The rework reduction alone pays for the investment.
  • Track cycle time, throughput, work item age, and WIP. These four flow metrics predict delivery better than velocity ever could.