Email Automation Testing: How to Continuously Improve Your Flow Performance
Most brands set up their email automation flows, move on to campaigns and other channel activity, and revisit the flows only when something goes obviously wrong. The welcome series was built 18 months ago and has never been touched. The abandoned cart flow has the same subject lines it launched with. The post-purchase sequence has not been updated since the brand redesign.
This is the set-and-forget trap — and it is costing significant recoverable revenue.
Email automation flows are not static revenue generators. They are systems that respond to testing and optimisation the same way any marketing asset does. The difference is that improvements to flows compound indefinitely — every change that improves performance applies to every subscriber who enters that flow going forward. Unlike campaign improvements (which expire with each send), flow improvements are permanent returns on a single optimisation investment.
This guide covers the systematic testing framework for continuously improving automation flow performance — what to test, how to structure experiments in Klaviyo, how to interpret the data, and the compound improvement model that makes consistent testing the highest-ROI activity in email marketing.
Why Most Brands Never Test Their Flows
The reasons brands do not test automation flows are predictable and understandable.
First, flows feel done. A campaign is a temporary piece of work with a clear end date. A flow is infrastructure — it is built, it runs, and it generates revenue. The psychological sense that flows are “complete” discourages the ongoing optimisation mindset.
Second, the effort to set up a flow test in Klaviyo feels disproportionate to the apparent benefit. Brands that test subject lines in campaigns regularly still never run tests in their flows, simply because the workflow is less familiar.
Third, the results are not as immediately visible as campaign results. A campaign test shows results within 24 hours. A flow test might need 3–4 weeks to accumulate statistically meaningful volume, especially on lower-traffic flows.
None of these reasons reflects the actual economics. A 10% improvement in abandoned cart flow open rate generates incremental revenue every month for the foreseeable future. The payback period on the effort to set up and interpret a single flow test is typically measured in weeks.
The Testing Priority Hierarchy: What to Test First
Not every element of every flow is worth testing. Prioritise tests based on their potential revenue impact.
Priority 1: Subject Lines on High-Volume Flows
Subject lines have the largest direct impact on flow revenue because open rate is the first gate in the conversion funnel — subscribers who do not open cannot click, and subscribers who do not click cannot purchase.
Start with subject line tests on your highest-volume flows: abandoned cart (first email), welcome series (first email), and post-purchase (first email). These are the flows with the most traffic, which means tests reach statistical significance faster and the revenue impact of an improvement is largest.
A 10% improvement in open rate on an abandoned cart first email — say from 45% to 55% open rate — translates proportionally to more clicks, more conversions, and more revenue, every month.
Priority 2: Send Timing on Key Flow Steps
The timing between emails in a sequence significantly affects both engagement and the subscriber’s state of mind when they receive each email.
For abandoned cart, the first email timing is the most critical test. The industry benchmark is 1 hour after abandonment, but some brands see better results at 30 minutes (faster, higher urgency) or 2 hours (giving more browse time). Test your specific audience’s optimal timing.
For post-purchase, the timing of the cross-sell email relative to the estimated delivery date is a high-impact timing variable. Sending cross-sell content 3 days after delivery (when the subscriber has experienced the product) outperforms sending it 3 days after purchase (before delivery) for most product categories.
For welcome series, the gap between emails 1 and 2 affects dropout rate significantly. Test 1 day vs 2 days between the first and second email, and between each subsequent pair.
Priority 3: Discount vs No Discount
Many brands include a discount in their abandoned cart flow by default. The value of testing no-discount variants is significant both for revenue (higher margin on non-discounted conversions) and for customer behaviour (subscribers who learn that waiting for the third abandoned cart email produces a discount will start abandoning carts to get it).
Test a variant of your abandoned cart flow where email 3 has no discount, or where the discount is replaced with free shipping, with a gift wrap option, or with exclusive early access to the next sale. Measure conversion rate, not just click rate — the relevant question is whether the conversion loss (if any) is offset by the margin gain.
Priority 4: Email Count in a Sequence
The optimal number of emails in a sequence is not universal. A welcome series that works best as 4 emails for one brand might perform better as 6 for another. An abandoned cart sequence that needs 3 emails for a high-AOV product might need only 2 for a low-AOV impulse purchase.
Test email count by creating a variant flow with one fewer email (removing the lowest-performing email in the sequence) and comparing cumulative flow revenue per subscriber over 30 days.
Priority 5: Content Angle
Content angle tests are the most complex and time-intensive flow tests, but they can produce the largest improvements when the current content is not well-matched to subscriber psychology.
Content angle tests compare: feature-benefit copy vs storytelling copy, product-focused vs customer-story-focused, educational vs promotional, brand voice styles (warm and personal vs direct and efficient).
These tests require at least 4–6 weeks and significant sample sizes to reach reliable conclusions. Run them after subject line and timing tests have been completed on a given flow.
How to Run A/B Tests Within Automation Flows in Klaviyo
Klaviyo’s flow A/B testing feature allows you to create split tests at any point within a flow — testing subject lines, send times, content, or even entire email variations.
To set up a flow A/B test in Klaviyo:
Navigate to the flow you want to test and select the email you want to split test.
Click “Add A/B test” to create a split. By default, Klaviyo sends 50% of flow recipients to each variant, though you can adjust this split.
Define your variants. For a subject line test, variant A receives subject line A and variant B receives subject line B. Everything else is identical.
Set your winning metric. For most flow tests, this should be “Revenue per recipient” or “Conversion rate” rather than “Open rate” — you want to optimise for revenue, and open rate improvements do not always translate to revenue improvements if the email content does not support conversion.
Set a test duration. Klaviyo can automatically declare a winner after a specified period or volume threshold. For most flows, set a minimum of 2 weeks and a minimum sample size of 500 per variant before declaring a winner.
Once the test concludes, implement the winning variant as the default for that flow step, document the result in your testing log, and plan the next test.
Interpreting Flow Experiment Data
Flow test data requires more careful interpretation than campaign test data because of the time dimension.
A flow test that runs for 2 weeks is seeing a cohort of subscribers who entered the flow during that 2-week period. This cohort may be different from subscribers who entered in a different period — for example, if you ran the test during a sale period, the engagement context is different from normal periods. Note the context when documenting results.
Statistical significance in flow tests is often harder to achieve on low-volume flows. For flows that process fewer than 200 subscribers per month (which is true of many advanced flows), meaningful A/B tests are difficult. For these flows, focus on implementing known best practices rather than testing — reserve testing for your high-volume flows where significance is achievable.
When interpreting multi-metric results (variant A wins on open rate, variant B wins on revenue per recipient), prioritise the business metric: revenue per recipient is almost always the most important metric in a commercial email context.
The Compound Improvement Model
The compounding arithmetic of flow testing is the strongest argument for running an ongoing flow optimisation programme.
Assume you run one test per flow per quarter across 5 flows. That is 20 tests per year. Assume each test produces an average 5% improvement in revenue per recipient for the winning variant (a conservative estimate — well-designed tests often produce larger lifts).
In year one, the compounding works like this: the first improvement on each flow is 5%. By the second improvement on each flow, the baseline is already 5% higher, so the second 5% improvement is a 5% improvement on a higher base. Over four improvements per flow per year, the cumulative revenue per recipient improvement is approximately 21% per flow.
Across five flows, this represents a 21% improvement in automation revenue — without increasing list size, without more campaigns, and without significant ongoing creative investment beyond the test setup.
Over two years of consistent testing, the cumulative improvement is approximately 40–50% higher automation revenue versus the baseline. This is the compound improvement model — and it is the reason that brands with a systematic flow testing programme consistently outperform their benchmarks over time.
Building an Ongoing Flow Optimisation Calendar
A practical flow optimisation calendar ensures testing happens consistently rather than opportunistically.
Structure the calendar by flow and by quarter:
Q1 (January–March): Welcome series subject line tests, abandoned cart timing tests.
Q2 (April–June): Post-purchase sequence content angle tests, browse abandonment subject line tests.
Q3 (July–September): Abandoned cart discount vs no-discount tests, welcome series email count tests.
Q4 (October–December): BFCM period is not ideal for testing (too much contextual noise). Use Q4 to implement winning variants from Q1–Q3 and plan the following year’s testing programme.
Each test should be documented in a running log with: flow name, email tested, variant descriptions, start and end dates, sample sizes, metrics, winner, and notes on context. This log becomes your brand’s email optimisation knowledge base — informing decisions for years.
Flow testing is one of the highest-ROI activities in email marketing and one of the most consistently under-invested. The brands that build a systematic testing programme and stick to it for 12–24 months create compounding advantages that become very difficult for competitors to replicate.
At Excelohunt, we include flow testing as a core part of every retainer we run. The testing programme is not an add-on — it is how we continuously justify the investment.
Related Excelohunt Services
Looking to implement these strategies with expert support?
- Email Automations — learn how we implement this for clients
- A/B Testing — learn how we implement this for clients Book a free strategy call with Excelohunt →
Want Us to Implement This for Your Brand?
Get a free email audit and see exactly where you're losing revenue.
Get Your Free Audit