Experimentation Rituals That Actually Scale
Background
Running experiments is deceptively easy; running repeatable, trustworthy, and impactful experiments that teams can consistently rely on is a far greater challenge. Just a year ago, our growth squads were struggling to execute even one experiment per month. These experiments often suffered from a lack of clean, reliable data, making it difficult to draw definitive conclusions, and follow-up analysis was frequently an afterthought. This ad-hoc approach led to a significant amount of wasted effort, inconclusive results, and a general lack of confidence in our experimentation program.
Recognizing that a robust experimentation culture was critical to doubling our growth impact, we undertook a comprehensive overhaul of our entire process. Our strategy involved a deliberate blend of established agile rituals, the integration of strong, end-to-end telemetry, and a clear codification of ownership at every stage. By borrowing best practices from agile methodologies, we instilled a rhythm and discipline into our experimentation cycles. The introduction of robust telemetry ensured that every experiment was meticulously tracked, providing clean data for rigorous analysis. Furthermore, by clearly defining who was responsible for each aspect of an experiment, from ideation to debrief, we fostered accountability and streamlined execution.
The results of this transformation have been remarkable. We now consistently average four experiments per sprint, spanning across various surfaces including web, messaging, and automation. This post delves into the specific rituals, tools, and cultural shifts that enabled us to achieve this significant leap in our experimentation capabilities, transforming a chaotic process into a predictable engine for growth.
Operating Principles
- Hypotheses must be falsifiable and tied to a single metric.
- Every experiment has an explicit owner from ideation through debrief.
- Telemetry is non-negotiable. No tracking plan, no launch.
- Feature flags everywhere. Rollouts happen in minutes, not weeks.
- Document everything. Wins, neutral results, and failures live in the same library.
Tooling Overview
| Need | Tool | Notes | |------|------|-------| | Backlog & workflow | Linear | We utilize Linear for managing our experimentation backlog and workflow. A dedicated "EX" team within Linear uses standardized templates for experiment intake, quality assurance (QA), and debriefs, ensuring a consistent and organized process from start to finish. | | Telemetry | PostHog + dbt | PostHog serves as our primary event capture and analytics platform. Coupled with dbt (data build tool), we've established a shared metrics layer that provides robust guardrails for sample size calculations and statistical significance, ensuring data integrity and reliable analysis. | | Feature flags | PostHog flags + custom middleware | For dynamic control over experiment rollouts, we leverage PostHog's feature flagging capabilities. This is augmented with custom middleware that provides granular control, offering a level of flexibility and power comparable to enterprise solutions like LaunchDarkly, but without the associated cost. | | Design review | Figma + Storybook | Our design process integrates Figma for UI/UX design and Storybook for component development. This ensures that all UI changes are meticulously reviewed and that components are synced with our design tokens, guaranteeing that visual changes in experiments accurately mirror the product's established design system. | | Communication | Resend summaries + #experiments Slack channel | Effective communication is vital. We use Resend to automate weekly experiment recaps, which are then shared in a dedicated #experiments Slack channel. This keeps all stakeholders informed about ongoing experiments, key learnings, and upcoming initiatives. |
Prioritization Rubric
We grade every proposal against the FICE rubric (Fit, Impact, Confidence, Effort). Each dimension is scored 1â5 with explicit prompts so debates stay objective:
- Fit. Does the experiment ladder up to the quarterly mission and the owning squadâs KPIs? We reject clever one-offs that donât map to strategy.
- Impact. We estimate the size of the population touched and the expected lift using historical conversion bands. Analysts provide the model, not opinions.
- Confidence. Do we have directional evidence (qual insights, prior tests, user interviews) that justify the bet? High-confidence scores require linked artifacts.
- Effort. Engineering, design, and data sign off on the engineering hours plus QA cost. Anything above 5 days needs director approval.
The rubric lives directly in the Linear intake form so submitters know how proposals will be judged. It also makes trade-offs transparent when stakeholders ask why their idea waited a sprint.
Key Principles of Scalable Experimentation
- Hypothesis-driven approach: Every experiment must start with a clear, falsifiable hypothesis tied to a single, measurable metric.
- Dedicated ownership: Assign explicit owners for each experiment from ideation through debrief to ensure accountability and continuity.
- Robust telemetry: Implement comprehensive tracking plans and a shared metrics layer to guarantee clean, reliable data for analysis.
- Ubiquitous feature flagging: Utilize feature flags for rapid, controlled rollouts and instant rollbacks, minimizing risk and maximizing agility.
- Thorough documentation: Maintain a centralized library of all experiments, including wins, neutral results, and failures, to foster organizational learning.
- Automated quality assurance: Integrate automated tests for visual regressions, accessibility, and data integrity into the experimentation pipeline.
- Continuous feedback loops: Establish regular rituals for reviewing results, sharing learnings, and refining the experimentation process.
Rituals in Detail
Weekly Intake (Monday)
Growth, product, and engineering meet for 30 minutes to review experiment proposals submitted via Linear. Each submission includes hypothesis, target metric, guardrail metrics, rollout plan, and expected engineering effort. We accept, defer, or reject proposals on the spot.
Sprint Planning (Tuesday)
Accepted experiments receive:
- Assigned engineer + analyst.
- Design assets or copy requirements.
- Tracking plan (Segment events + dbt model updates).
- QA checklist covering browsers, devices, and accessibility.
We limit ourselves to four concurrent experiments per surface to avoid noise.
Daily Standups
Each experiment owner gives a 60-second update: implementation status, blockers, telemetry readiness. If a blocker requires cross-team help, we escalate immediately rather than waiting for retro.
Release + QA
- Feature flag created with default OFF.
- Playwright regression tests run automatically.
- Analytics QA: we use a custom script that fires PageView + custom events, then verifies they landed in PostHog/dbt within five minutes.
- Launch staged to 5% of traffic; health monitored for 30 minutes before scaling.
Instrumentation & QA Checklist
Every experiment has a living checklist that must turn green before launch:
- Events defined. Segment + PostHog schema reviewed by analytics, including naming conventions and guardrail metrics.
- Data contracts verified. dbt tests run in CI to ensure new events donât break downstream models.
- Accessibility + performance pass. Lighthouse and axe CI checks must be â„90; perf regressions require a compensating experiment.
- Rollback plan documented. For each flag we sketch âwhat happens if we revert in the middle of onboardingâ so CS and Sales arenât surprised.
- Owner escalation tree. If metrics swing beyond guardrails, everyone knows who pulls the plug and who communicates with leadership.
It sounds heavy, but the checklist sits next to the PR and takes five minutes to verify because the automation does most of the work.
Friday Demo + Retro
We demo results (even if still running) and log learnings into Notion. Each entry includes:
- Link to code/flag.
- Screenshots/videos.
- Metric movement with confidence intervals.
- Decision (ship, iterate, revert, archive).
Sample Timeline
| Day | Activity | |-----|----------| | Monday | Intake + prioritization | | Tuesday | Planning + instrumentation reviews | | Wednesday | Build + QA | | Thursday | Launch + monitoring | | Friday | Demo + retro + backlog grooming |
Org Enablement and Incentives
Process dies when incentives clash. We invited marketing, success, and sales ops to nominate âexperiment buddiesâ who shadow a squad for two sprints. Buddies help write hypotheses, sanity-check guardrails, and carry learnings back to their teams. Leadership reviews a one-page âexperiments shipped vs. learnings capturedâ scorecard in the weekly business review, so executives celebrate disciplined invalidations rather than only wins. We also track experiment debt (tests stuck without analysis) and treat it like tech debtâif the number creeps above three, we pause new launches until analysis catches up.
Scaling to Multiple Squads
Once two squads adopted the rituals, we packaged them into a starter kit:
- Notion workspace template with linked databases for hypotheses, rollouts, and retros.
- Linear project template that auto-generates the right statuses, custom fields, and automation rules.
- Starter telemetry pack (dbt macros + dashboard) so new squads donât reinvent charts.
- Buddy rotation calendar assigning an analyst and designer to each new squad for their first sprint.
We launch new squads in âmentored modeâ for two iterations. After that they enter the shared operating rhythm and contribute to the weekly backlog refinement.
Metrics We Track
- Experiments launched per sprint: We consistently launch an average of 4 experiments per sprint, aligning with our goal of 3â5.
- Analysis completion rate: Our analysis completion rate is 100%, ensuring no open loops or unanalyzed experiments.
- Average time to decision: The average time from experiment launch to a clear decision (ship, iterate, revert, archive) is 6.2 days.
- Experiment win rate: We maintain an experiment win rate of 27%, with the remaining experiments providing either neutral or instructive failure insights.
- Technical rollback rate: Our technical rollback rate is consistently below 2%, a testament to the reliability of our feature flags and rigorous QA processes.
Templates and Documentation
Our Linear template includes sections for hypothesis, design artifacts, tracking plan, QA checklist, flag configuration, and postmortem summary. Notion serves as an indexed knowledge base; we tag experiments by surface, metric, and outcome so future teams can learn quickly. Resend automates weekly digests summarizing what shipped, whatâs mid-flight, and what we learned.
Lessons Learned
- Enforce guardrails. Sample-size calculators and stopping rules are embedded into our dashboards. If an experiment tries to end early, the analyst gets paged.
- Tie experiments to north-star metrics. Vanity tests are an easy trap; we require explicit links to OKRs.
- Share failures loudly. The best ideas often reseed from what didnât work. Transparency keeps the culture healthy.
- Make QA everyoneâs job. Designers review visuals, analysts verify events, engineers ensure performance, PMs sanity-check UX.
FAQ
How do you avoid experiment collisions?
We maintain a shared calendar inside Notion showing which cohorts are exposed to which flags. PostHog cohorts are tagged with experiment IDs, and middleware blocks conflicting flags from activating simultaneously.
What about long-running experiments?
Anything exceeding three sprints requires a special review where we revalidate assumptions and consider alternative designs (like staggered rollouts or synthetic control groups).
Do stakeholders outside engineering access results?
Yes. Resend digests include executive summaries plus links to dashboards. We also host a monthly âgrowth show-and-tellâ where product marketing, success, and sales learn from experiment outcomes.
Whatâs Next
Weâre piloting auto-generated hypotheses using telemetry + LLMs to help teams spot underexplored segments. Weâre also adding automated debrief summaries so insights land in Slack minutes after an experiment wraps. If you want the Linear template, tracking plan checklist, or Resend digest snippet, reach out and weâll send the package.