Introduction: High Velocity ≠ Good Delivery
Your last sprint clocked 80 story points. This one: 52. The sprint review is tense. Everyone's looking for explanations—was it that pipeline outage? Another schema change in the upstream data source?
But what if I told you that the 52-point sprint actually delivered better data quality to downstream users than the 80-point one? Would you still care about Velocity?
Data engineering teams have a widespread cognitive trap: copy-pasting software Agile practices wholesale and using Velocity as the primary measure of team productivity. This isn't just ineffective—it can be actively harmful.
Why Velocity Distorts in Data Teams
| Symptom | Root Cause |
|---|---|
| Velocity crashes in a sprint | Emergency fix from upstream schema change |
| Velocity spikes unrealistically | "Investigation tasks" counted as story points—zero actual deliverables |
| Story point estimates consistently wrong | Business definitions change mid-implementation |
| Tasks "not done" at sprint end | Data quality tests failing, pipelines in limbo |
The core mismatch: traditional Velocity was designed for feature delivery, but data work is an inherently exploratory + corrective blend that resists neat framing.
Three Layers of Metrics That Actually Matter
Layer 1: Core Scrum Metrics (Required)
| Metric | Definition | How Data Teams Should Use It |
|---|---|---|
| Sprint Velocity | Total SPs completed per sprint | Use a 4–6 sprint rolling average. Never compare single sprints. |
| Sprint Goal Achievement | Did the sprint's core objective get met? | 10x more important than Velocity. Frame goals around "data available," not "task completed." |
| Carry-over Rate | Unfinished tasks rolled to next sprint | Over 20% is a red flag—signals estimation or dependency management problems. |
Layer 2: Data Quality Metrics (Data Team Exclusive)
- Pipeline SLA Achievement Rate: Are critical pipelines producing data by their agreed deadlines?
- Data Downtime: Percentage of time downstream users can't access data due to pipeline failures
- P0 Bug MTTR: Mean time to resolve data production incidents
- Data Quality Test Pass Rate: Pass rate of automated tests (dbt tests, Great Expectations, etc.)
Layer 3: Outcome Metrics (The Layer Everyone Forgets)
- Downstream User Activity: Who uses the data we produce? Is usage trending up or down?
- Self-Service Coverage Rate: What percentage of business questions can be answered with existing datasets (without ad-hoc requests)?
- Backlog Trend: Is the queue of pending data requests growing or shrinking?
High Velocity + Low Outcome Metrics = doing the wrong things correctly.
Sprint Capacity Planning for Data Teams
The "Protected Capacity" Principle
Explicitly reserve maintenance capacity in every sprint—never allocate 100% to new development:
Planned new feature development: 60–70%
Pipeline maintenance/bug buffer: 15–20%
Spikes/technical research: 10–15%
Retrospective/docs/code review: 5%
Maintenance load is consistently underestimated. Without explicit reservation, data teams almost always get buried by surprise incidents mid-sprint.
How to Write Sprint Goals That Work
❌ Bad goal: "Complete the ETL refactor for Sales Pipeline, migrate to dbt"
✅ Good goal: "Sales team can see real-time month-to-date deal data in the dashboard (latency < 1 hour)"
Good goals make it clear who benefits from the sprint—and give you an unambiguous success criterion for the retrospective.
Five Anti-Patterns to Avoid
- Velocity competitions: Evaluating teams by Velocity incentivizes inflating story point estimates (points inflation), hiding real problems.
- Unlimited mid-sprint ad-hoc requests: When everyone says "this is urgent," Sprint Goals become meaningless. Set a "sprint lock" period.
- Quantity without quality: If the pipeline runs but data is wrong—is that Done? Data quality tests must be part of your DoD (Definition of Done).
- Skipping retrospectives: Data teams over-index on technical work and cut Retros. But Retros are the mechanism for surfacing systemic issues (recurring pipeline failures, chronic estimation drift).
- Kanban-vs-Scrum binary thinking: Mature data teams run a "dual-track" model—Scrum for planned platform work, a Kanban lane for urgent ad-hoc requests. They don't conflict; they complement.
Conclusion: Track Meaningful Things
The trap data teams fall into most isn't technical debt—it's metrics debt: driving the team with wrong indicators, then wondering why quality doesn't improve.
A proper metrics framework needs three layers running simultaneously:
- Scrum fundamentals (stability)
- Data quality indicators (reliability)
- Outcome metrics (effectiveness)
These three layers aren't overhead. They're a navigation system that tells you how fast you're moving, in what direction, and whether you're actually getting somewhere.
A data team's maturity isn't measured by which tools they use. It's measured by whether they know whose work they're creating value for—and how much.
Sources: monday.com Agile Best Practices | Medium: Agile in Data Teams | Lonti: Agile Data Management
Top comments (0)