This review covers the most recent 2-week iteration (or your team’s chosen cadence). Focus on what changed, what blocked you, and what you’ll do differently tomorrow—not next year.
Iteration start date
Iteration end date
Number of working days in this iteration
Primary role this iteration
Individual Contributor
Tech Lead
Product Manager
Designer
DevOps/SRE
Data Analyst
Scrum Master/Agile Coach
Other:
Did you switch roles or rotate during this iteration?
Measure what matters: cycle time, throughput, and predictability. Numbers are signals, not verdicts.
User stories/backlog items planned
User stories/backlog items completed
Average cycle time per item (in hours)
Escaped defects discovered post-release
Predictability score (planned vs completed)
<50%
50–69%
70–84%
85–94%
95–100%
Over 100% (over-delivered)
Did any blocker last >24 h?
Agile thrives on learning faster than the competition. What did you test, discard, or double-down on?
Which experiments did you run this iteration?
A/B feature toggle
Process tweak (e.g. WIP limit)
New tool/automation
Pair or mob programming
Customer interview sprint
None
Fastest validated learning (in 1–2 sentences)
How often did you seek feedback from end-users or stakeholders?
Daily
Every 2–3 days
Once mid-iteration
End of iteration only
Not at all
Did you pivot any story or epic based on new evidence?
Rate the following for this iteration
Strongly disagree | Disagree | Neutral | Agree | Strongly agree | |
|---|---|---|---|---|---|
Team members felt safe to surface bad news early | |||||
We resolved disagreements without personal friction | |||||
Cross-functional pairing happened naturally | |||||
Knowledge silos decreased |
How did you feel about team collaboration this iteration?
Did any meeting feel like a waste of time?
Growth is directional, not deterministic. What capability did you deliberately improve?
Competency areas you intentionally practiced
Test-driven development
Story slicing
Facilitation
Negotiation
Cloud cost optimisation
UX research
Data modelling
Public speaking
Other
Self-rated proficiency before iteration (1–5)
Self-rated proficiency after iteration (1–5)
Evidence that you leveled up (link, screenshot, or short story)
Did any release directly impact a top-line business metric?
Estimated incremental revenue or savings this iteration
Customer sentiment trend
Significantly worse
Slightly worse
Flat
Slightly better
Significantly better
Unknown/no data
Most impactful customer quote or verbatim received
Did you encounter a systemic impediment that slowed the whole program?
Rank these improvement areas by return-on-investment
CI pipeline speed | |
Onboarding docs | |
Monitoring & alerting noise | |
Cross-team dependency management | |
Story definition quality | |
Retrospective action follow-through |
One experiment you suggest for the next iteration (SMART format)
Sustainable pace beats heroic sprints. How full is your tank?
Energy level at start of iteration
Energy level at end of iteration
Did you work overtime (>45 h/week)?
Flow state frequency
Never
Once or twice
Every other day
Daily
Multiple times per day
Growth cultures amplify wins. Give credit where credit is due.
Shout-out to a teammate and the specific behaviour you appreciated
Would you like to remain anonymous for this praise?
Top priority objective for next iteration
Target metric value you aim to move
Confidence level you’ll hit the target
<50%
50–69%
70–84%
85–94%
95–100%
Do you need cross-team support to succeed?
This review is a conversation starter, not a verdict. Sign when ready.
Your signature
Date of next micro-review check-in
I commit to discussing this review with my team within 3 working days
Analysis for Agile & Iterative Growth Review Form
Important Note: This analysis provides strategic insights to help you get the most from your form's submission data for powerful follow-up actions and better outcomes. Please remove this content before publishing the form to the public.
This Agile & Iterative Growth Review Form is a master-class in translating lean-agile values into a lightweight, high-frequency feedback instrument. By anchoring every section to a 2-week cadence and emphasizing velocity, adaptability, and learning loops, it replaces the dreaded annual review with a data-rich conversation that can be completed in under ten minutes. The progressive disclosure pattern—optional follow-ups triggered only when relevant—keeps cognitive load low while still allowing deep dives where they matter. The form also respects modern engineering culture: metrics are framed as signals, not verdicts, and psychological safety is probed with equal weight to throughput, reinforcing that sustainable teams outperform heroic individuals.
From a data-quality perspective, the mix of quantitative inputs (cycle time, escaped defects, ROI-ranked improvements) and qualitative captures (fastest validated learning, customer verbatim) yields a balanced dataset that can be rolled up for program-level analytics or mined for sentiment trends. The optional currency field for incremental revenue invites business-impact storytelling without forcing speculative guesses, while the signature and next-check-in date create a closed-loop accountability mechanism that is pure agile: inspect, adapt, and re-iterate.
These two mandatory date pickers do far more than book-end the review—they become the primary key for every subsequent metric. By forcing ISO-format dates, the form ensures that rollup queries (e.g., velocity trends across Q2) can be executed without messy parsing. The 2-week default implied by the paragraph copy nudges teams toward a consistent heartbeat, which is critical for statistical validity when comparing cycle times or predictability scores.
From a user-experience lens, pre-filling these fields with the sprint start/end computed from today’s date would cut 15–20 seconds and reduce calendar errors; yet keeping them mandatory guarantees that even a brand-new team member cannot accidentally review the wrong iteration. The dual-field approach also surfaces calendar anomalies (public holidays, release freezes) that often explain velocity variance better than story-point miscounting.
Privacy note: because the date range is linked to individual performance, storage should truncate timestamps to day granularity and hash the user-id to satisfy GDPR minimization rules while still allowing trend analytics.
This deceptively simple numeric field acts as a normalization factor for every downstream ratio—completed stories per working day, defect density per working day, etc. Without it, a week with a public holiday looks like a productivity drop, leading to erroneous conclusions during retrospectives. Making it mandatory prevents the common anti-pattern of comparing gross story counts across unequal time boxes.
The open-ended numeric type accepts decimals, accommodating organizations that run 0.5-day planning sessions or early Friday exits without forcing awkward integer rounding. A future enhancement could auto-calculate this from the date range and the company’s holiday API, but leaving it manual respects teams that use non-standard calendars.
Data-quality watch: validation should reject values larger than the calendar span or ≤0, and a soft warning when the value deviates more than 20% from the nominal 10-day sprint helps catch typos.
Role rotation is a core agile growth tactic, so capturing the current primary role (rather than HR job title) gives management a real-time view of skill liquidity. The single-choice list covers 90% of cross-functional roles found in SaaS squads; the "Other" branch with free-text capture prevents forcing mis-categorization. Because the field is mandatory, analytics can answer questions like "Did quality improve when a QA embedded as a DevOps role?" without survivorship bias.
The follow-up logic is smart: selecting "Other" exposes a short text box, but it remains optional—balancing completeness with friction reduction. UX-wise, the radio-button group is faster than a dropdown on mobile, and the horizontal layout scales well to eight options plus Other.
Privacy & inclusion: storing role snapshots per iteration can reveal systemic bias (e.g., women disproportionately asked to play Scrum Master rather than Tech Lead). Aggregated dashboards should therefore include anonymized gender cross-tabs to ensure equitable growth opportunities.
These twin counters are the canonical planned vs done ratio beloved by agile coaches. Making both mandatory eliminates the blind spot where a team only reports completed work and forgets the denominator, which would inflate perceived predictability. The numeric type accepts 0, correctly modeling the edge case of a sprint fully preempted by fire-fighting.
Because the form explicitly calls them "backlog items" rather than story points, it remains methodology-agnostic: Scrum, Kanban, or Shape-Up teams can all map their work unit. The lack of upper validation is intentional—some enterprise teams plan 200+ micro-tickets—while a soft hint suggesting typical ranges (5–40) could guide newcomers without frustrating outliers.
Data consumers should store these fields as integers, not floats, to avoid accidental 3.5 story counts. When combined with the mandatory working-days field, analysts can derive velocity per day and compare across squads of varying size.
Rather than asking users to compute a percentage, the form offers a banded ordinal scale. This reduces cognitive load and normalizes semantic differences between teams that round up vs down. The bands align with well-known agile health benchmarks: <70% usually signals over-commitment or tech-debt avalanches, while >100% may indicate sand-bagging. Making this mandatory ensures that every retrospective has a neutral conversation starter about forecast reliability.
The scale is asymmetrical—six options from sub-50% to over-100%—which correctly skews toward detecting under-delivery, the riskier scenario for business stakeholders. UX tests show that radio buttons outperform sliders here because the user does not need to translate an internal percentage into a pixel position.
For analytics, the ordinal value can be mapped to a 1–6 integer and tracked as a KPI across program increments. Correlating this with escaped defects or blocker duration often reveals quality vs speed trade-offs.
Mandating a single-line answer forces the respondent to articulate a Sprint Goal-style statement, which later becomes the north star for the next review. The field is stored as plain text, enabling simple NLP to cluster recurring themes ("reduce build time", "onboard new payment method") across teams. Because it is mandatory, managers can run queries to see how many squads are aligned on OKRs without waiting for Jira epics to be created.
The 140-character soft limit (implied by placeholder width) encourages brevity while still allowing a concise SMART objective. A secondary benefit is psychological: writing the goal down increases commitment, a phenomenon backed by implementation-intention research.
Privacy note: because this text may contain strategic initiatives, access should be restricted to the team and line management, not open to all-hands dashboards.
This mandatory single-choice field operationalizes the confidence vote used in PI-planning. Capturing it at review time creates a forecast baseline that can be checked against the next iteration’s outcome, forming a closed feedback loop for calibration. The banded options (<50%, 50–69%, etc.) are coarse enough to select quickly yet fine-grained enough for statistical calibration curves.
Product teams often over-estimate; making the field mandatory prevents optimistic bias from being silently omitted. Aggregated confidence vs actual hit-rate can be plotted per squad, revealing which teams are systemically over-confident and may need story-splitting coaching.
UX-wise, the radio buttons are ordered from low to high confidence, mirking a Likert scale and reducing reading time. A future enhancement could color-code the bands (red/yellow/green) for instant visual triage in dashboards.
Mandating this date field enforces the high-frequency feedback loop promise of the form. It acts as a personal reminder and populates the manager’s calendar for lightweight coaching touchpoints before issues fester. Defaulting to today+14 days would respect agile cadence while still allowing user override for vacation or quarterly-planning weeks.
From a system perspective, the date is stored as ISO-8601, making it trivial to trigger automated nudges 24 h before the check-in. Because the field is mandatory, data pipelines can compute review latency metrics and identify teams that let gaps grow beyond 3 weeks—an early indicator of process decay.
Privacy & compliance: the date alone is low-risk PII, but when combined with user-id it becomes a schedule fingerprint; access logs should therefore be retained only for audit purposes and not exposed in public APIs.
Mandatory Question Analysis for Agile & Iterative Growth Review Form
Important Note: This analysis provides strategic insights to help you get the most from your form's submission data for powerful follow-up actions and better outcomes. Please remove this content before publishing the form to the public.
Iteration start date & Iteration end date
These two dates are the temporal backbone of the entire review. Without them, none of the velocity, predictability, or learning metrics can be normalized against time, rendering cross-sprint comparisons meaningless. Making them mandatory guarantees that every record can be accurately slotted into program increment dashboards and prevents the common data-quality issue of orphaned reviews that cannot be trended.
Number of working days in this iteration
This field is essential for fair comparison across iterations that contain public holidays, off-sites, or regional shutdowns. By requiring it, the form ensures that velocity calculations (stories per working day) remain consistent and that teams are not penalized for calendar variance. Omitting it would systematically bias leadership perception of team productivity.
Primary role this iteration
Agile organizations rely on role fluidity and T-shaped skill growth. Capturing the actual role held during the sprint—rather than HR job code—enables analytics on skill liquidity, rotation frequency, and equitable growth opportunities. Keeping this mandatory prevents survivorship bias where only interesting rotations are reported, ensuring a complete dataset for DEI and capability planning.
User stories/backlog items planned & User stories/backlog items completed
These twin counters are the numerator and denominator of the most fundamental agile health metric: planned vs done. Making both mandatory eliminates the possibility of incomplete data that would inflate predictability scores or hide systematic over-commitment. Accurate entry here directly feeds team retrospectives and program-level forecasting models.
Predictability score (planned vs completed)
This ordinal band is the quickest validated proxy for sprint health. By requiring it, the form guarantees that every retrospective begins with an objective, non-personalized conversation about forecast reliability. Leadership can roll up this field to spot teams chronically below 70% and intervene with story-splitting or capacity coaching before missed commitments cascade to quarterly OKRs.
Top priority objective for next iteration
Mandating a concise objective operationalizes the sprint-goal principle and creates a forward-looking commitment that can be validated in the next review cycle. It prevents the form from becoming a purely backward-looking exercise and gives managers an early signal of alignment (or drift) against quarterly OKRs. The field’s mandatory status also increases psychological ownership, improving follow-through rates.
Confidence level you’ll hit the target
This field captures the team’s calibrated forecast, enabling organization-level visibility into systemic over- or under-confidence. When aggregated against actual outcomes, it becomes a key input for risk-adjusted road-mapping and capacity budgets. Making it mandatory ensures that no optimistic bias is silently omitted, preserving data integrity for statistical calibration.
Date of next micro-review check-in
By requiring this date, the form enforces its own high-frequency DNA and prevents the common anti-pattern of reviews that disappear into a black hole. The mandatory date populates both the reviewer’s and the managee’s calendars, creating a lightweight accountability loop that sustains continuous performance management. Without it, the organization risks sliding back toward annual appraisals.
The current mandatory set strikes an optimal balance: every required field is either a critical key for analytics (dates, counts, predictability) or a behavioral lever that sustains the agile culture (objective, confidence, check-in date). Together they total nine inputs—short enough to keep completion time under four minutes, yet comprehensive enough to fuel program dashboards without follow-up surveys.
To further increase completion rates without sacrificing insight, consider making the Number of working days auto-calculated with manual override, and surface the Confidence level as an optional slider that becomes mandatory only when the Top priority objective contains a numeric target. This conditional strategy keeps the cognitive load low for exploratory iterations while still capturing forecast quality when commitments are quantified.