methodology

mini-cascades. a way of building software with humans and agents in the loop.

steerdev's methodology is a small set of opinions about how software gets shipped when implementation is no longer the bottleneck. it is not a framework, a certification, or a religion. it is a working draft — the shape that produced good results for us, written down so others can try it, push back, and improve it.

the bottleneck moved. the methods didn't.

most of the rituals on a modern engineering team — the standups, the story-pointing, the sprint cadences, the line-by-line code reviews — were designed for a world where writing the code was the slow, expensive step. that world is fading. capable agents now write usable code in minutes. the new slow steps are different ones: deciding what to build, decomposing it cleanly, and proving it works.

we tried bolting agents into the old rituals. it didn't work. PRs got bigger. reviews got worse. specs stayed in slack. agents shipped confidently wrong code that humans rubber-stamped because there was too much of it. the rituals were optimizing for the wrong thing.

mini-cascades are our attempt at rituals optimized for the new bottleneck — judgment over typing.

before AI
planning
less effort
implementation
more effort
verification
moderate effort
after AI
planning
more effort
implementation
less effort
verification
more effort

writing code was never the real bottleneck. understanding what to build and proving it works were always the hard parts. AI shifted where the effort lands — less of it on typing, more of it on the parts that always mattered. generate ten times the code and you need ten times the clarity upfront, ten times the rigor afterward.

three converging, one diverging, one converging again.

a mini-cascade is one trip through these five phases. iterations are scoped by risk, not the calendar — small low-risk changes can run a whole cycle in an afternoon; big architectural shifts get the time they need. the shape, though, is always the same.

01defineconverge ⊙

agree on what to build, before anyone writes anything.

a PRD captures intent — a problem, a constraint, a definition of done. it's authored by humans, often with an agent helping draft. the artifact is small and reviewed; it is not a placeholder for a slack thread.

02planconverge ⊙

expand intent into feature specs.

each problem decomposes into one or more feature specs: bounded units of behavior with explicit interfaces and known unknowns. AI helps draft and pressure-test. humans approve. the cost of a bad spec compounds; we pay it here, on purpose.

03decomposeconverge ⊙

turn feature specs into implementation specs and task waves.

features become implementation specs (the contract for the change) and tasks (the work). tasks declare their dependencies on each other. independent tasks run as a wave. blocked work is visible, not pretended-around.

04builddiverge ↯

the only diverging phase. parallel agents execute.

this is the only phase where breadth matters more than alignment. an agent picks up a task, executes against the implementation spec, produces evidence. multiple agents work the wave in parallel. they don't need to agree on style or approach — only on the contract their task declared.

05verifyconverge ⊙

collapse parallel work into a reviewable PR with evidence.

the cycle closes here. evidence is assembled: tests, screenshots, traces, the agent's own narration of decisions. humans review the evidence, not the diff. low-risk waves get a quick approval; high-risk ones get an explicit verification plan written before phase 04 even started.

what one cascade actually looks like. ten working days, day by day.

the shape is universal but the rhythm matters. here's how a typical cascade unfolds for a team of four engineers, a product manager, and a product owner. roughly forty-five percent of the cycle goes to planning, thirty to implementation, twenty-five to verification — the opposite of how most agile teams allocate time.

10working days
~45%on planning (P1–P3)
~30%on build (P4)
~25%on verify (P5)
dayphasefocuswho's driving
d1w1kickoff
P1
problem framing, business case, success metrics
pm + po lead, whole team
d2w1PRD draft
P1
scope in / scope out, edge cases, structured Q&A
pm authors, agents draft sections
d3w1PRD lock
P1
sign-off before any spec gets written
whole team review, ~2h
d4w1feature specs
P2
acceptance criteria, interfaces, entropy tolerance
engineers lead, AI drafts
d5w1spec refine
P2
2–3 iteration rounds, then approve
engineers + pm review
d6w2task waves
P3
decompose specs into wave-organized tasks
lead engineer + AI assist
d7w2wave 1 build
P4
devs steer; spec is the coordination mechanism
8–12 agents in parallel
d8w2wave 2 build
P4
unblocked tasks pick up as wave 1 lands
6–10 agents in parallel
d9w2verify
P5
evidence review, drift detection, integration tests
devs + pm + po
d10w2merge & ship
P5
staging → prod; retro begins next cascade
lead engineer deploys

two notes. first: the cadence is not sacred. low-risk changes can run the whole loop in an afternoon; an architectural shift might take three weeks. the shape stays. second: the build phase is the only one where humans aren't doing the typing. that's the point.

we review specs hard. we focus on the code that matters.

traditional teams review code obsessively and treat specs as a formality. when an agent produces eight hundred lines in twenty minutes, that ratio stops working. we redistribute the effort: significant review happens before any code exists, we read core and high-risk code closely, and the rest is reviewed through evidence — tests, traces, screenshots.

traditional · agile/scrum

most effort lands at the end, after the code is written
PRD review
5 min
spec review
15 min
task review
code review
2+ hours
QA cycle
1 day

mini-cascade · steerdev

effort spread across the cycle, weighted to where it pays off
PRD review
2 hours, whole team
spec review
1 hour, engineers
task review
30 min
evidence review
tests, traces, screenshots
code review
core code, in-depth

cascades overlap. the team is always shipping while always planning the next thing.

a single cascade is a closed loop, but a team running cascades back-to-back would idle half the staff during build phases and the other half during planning. so cascades pipeline. while cascade A is in build, cascade B starts its PRD. while A verifies, B is writing specs. the org runs two to three cascades in flight at any time, staggered so the work never bunches.

cascade A
P1P2P3P4P5
cascade B
P1P2P3P4P5
cascade C
P1P2P3P4P5
week 1week 2week 3week 4week 5

the practical effect: the team is always shipping while always planning. no two-week ceremony stalls. no end-of-sprint crunch. the engineer who finished steering wave two on cascade A spends day ten writing the PRD for cascade C.

not anti-agile. post-agile.

we keep what worked: short cycles, iterative improvement, frequent shipping, retros that change behaviour. we drop what doesn't survive contact with agents in the loop: ceremony overhead, synchronous coordination as the primary mechanism, the article-of-faith that documentation is bureaucratic drag.

dimensionagile / scrummini-cascades
cycle length
2-week sprints, fixed
2–3 week cascades, risk-driven
planning
light — stories, points, vibes~10–15% of time
heavy — PRDs, specs, task waves~45% of time
implementation
human-driven, AI-assisted
AI-driven, human-steered
ceremonies
standup · planning · review · retro · grooming~12 hrs / sprint
one front-loaded planning block, one verify blockno daily standup
coordination
synchronous (meetings)
structured context (shared specs)
verification
manual diff-by-diff code review
evidence-based — tests, traces, screenshots
documentation
minimal ("working software over docs")
central — the spec is the api between humans and agents
optimised for
human teams under human constraints
human–agent teams under AI economics

the question isn't whether agile dies. it's whether what replaces it still deserves the name. when sprint planning becomes intent design, standups become shared agent context, and code review becomes evidence review — that's a different methodology with the old branding.

seven commitments. not a framework, just the load-bearing ones.

01

specs are first-class artifacts.

PRDs, feature specs, and implementation specs are versioned and linked, not pasted into slack. agents read them. humans review them. they are the substrate the rest of the work runs on.

02

agents diverge. humans converge.

parallelism happens during build. judgment — what to build, whether it works — stays with humans. structure the workflow so neither party is doing the other's job.

03

risk sets the cadence.

a copy tweak ships in two hours. a payments refactor takes a week. let the work's risk drive the iteration length, not a fixed sprint.

04

prefer reviewing evidence, not diffs.

when an agent produces 800 lines of code, auditing every line breaks down. the agent must produce evidence — tests, traces, screenshots, an explanation. lean on that, while still reading core and high-risk code closely.

05

be agent-agnostic on purpose.

the agent landscape will keep shifting. the methodology shouldn't. structure your workflow so swapping agent vendors is a configuration change, not a re-architecture.

06

the spec is the api between humans and agents.

when communication is fuzzy, agents make confident wrong decisions. specs make the contract explicit. write them well; revise them when they fail.

07

agents should be proactive, not reactive.

if a human has to command every step, the human becomes the bottleneck. proactive agents take initiative — pick up the next task, run verification, surface evidence — guided by workflows and tools that let them act whenever they can.

things we used to do. things we stopped doing.

treating an agent like a faster typist.

if you're prompting an agent the same way you'd ask an intern to bang out a feature, you're wasting most of its leverage and absorbing all of its risk. specs first.

one giant PR with 4,000 lines of "AI did it".

no human can review that meaningfully. either decompose into smaller waves with their own evidence, or accept that "review" became "trust".

fixed two-week sprints for AI-driven work.

sprints calibrate humans against humans. with agents in the loop, batch sizes are wrong by an order of magnitude in either direction. let risk drive the cycle.

one "agent of record" lock-in.

the model that's best today probably isn't in six months. methodology should outlast any specific vendor.

PRDs as slack threads.

if the spec lives in chat, the agent can't read it, the next engineer can't find it, and the conversation is the artifact. unsustainable.

we don't have all the answers.

this methodology has produced good outcomes on real work. it has also raised questions we haven't fully answered. a few we're still chewing on:

  • how big should a wave be? we have heuristics, not a formula. teams running larger waves (10+ tasks) report different failure modes than smaller ones (3–5).
  • what's the right cadence for refactoring an agent's output? agents leave a particular flavor of inelegance. we're learning when to refactor inline vs. accept the debt vs. discard the work entirely.
  • how do you onboard humans into this? engineers joining a steerdev team don't have a 20-year tradition of tutorials to lean on. the apprenticeship is being figured out cycle by cycle.
  • what does this look like beyond software? we suspect the same shape — converging specs, diverging execution, converging evidence — generalizes. we haven't proven it.

if you have answers, opinions, or war stories, we'd love to hear them. this page will be revised as we learn — that's part of the methodology.