Strive Commerce
All Guides

Analytics & Data

A/B Testing for E-Commerce: A Practical Guide to Testing What Works

Master the fundamentals of A/B testing for your online store — from choosing what to test, setting up experiments correctly, reaching statistical significance, and avoiding common pitfalls.

10 min read

What Is A/B Testing?

A/B testing (also called split testing) is the process of comparing two versions of a web page or element to see which one performs better. You show Version A to half your visitors and Version B to the other half, then measure which version produces more conversions.

It is the scientific method applied to your store. Instead of guessing what works, you test it with real traffic and let the data decide.

Why A/B Testing Matters

Every opinion about what works best is just a hypothesis until tested. What you think customers want and what they actually respond to are often very different things.

Consider this: a store owner was convinced that a green add-to-cart button would convert better than an orange one because green means go. They tested it. The orange button won by 14%. Without the test, they would have left money on the table based on a wrong assumption.

A/B testing removes guesswork and replaces it with evidence.

What to Test (Prioritized)

Not all tests have equal impact. Focus on high-impact elements first:

High Impact

  • Headlines and value propositions: The first thing visitors read. Small wording changes can produce 20%+ conversion differences.
  • Call-to-action buttons: Text, color, size, and placement of your Add to Cart or Buy Now button.
  • Product images: Main product image, number of images, lifestyle vs. studio shots.
  • Price presentation: How you display the price, compare-at pricing, payment installment messaging.
  • Social proof placement: Where reviews and testimonials appear on the page.

Medium Impact

  • Product description length and format: Bullets vs. paragraphs, feature-focused vs. benefit-focused.
  • Page layout: Single column vs. two column, image placement relative to text.
  • Trust badges: Types, placement, and number of trust indicators.
  • Shipping information: Free shipping messaging, delivery time display.

Lower Impact

  • Font choices and sizes
  • Color schemes (beyond CTA buttons)
  • Footer content
  • Minor copy changes in non-critical areas

How to Run an A/B Test

Step 1: Choose One Variable

Test only one element at a time. If you change the headline AND the button color AND the image simultaneously, you cannot know which change caused any difference in performance.

Step 2: Define Your Metric

Decide what you are measuring before you start. Common metrics:

  • Conversion rate (purchases / sessions)
  • Add-to-cart rate (add-to-carts / sessions)
  • Click-through rate (clicks / views)
  • Revenue per visitor (total revenue / total visitors)

Pick one primary metric. You can track secondary metrics, but make your decision based on the primary one.

Step 3: Calculate Required Sample Size

This is where most people go wrong. You need enough traffic for statistically valid results.

Use a sample size calculator (Google "A/B test sample size calculator"). Input:

  • Your current conversion rate
  • The minimum detectable effect (how small a difference you want to detect)
  • Statistical significance level (use 95%)

For example, if your current conversion rate is 2% and you want to detect a 20% relative improvement (2% to 2.4%), you need approximately 10,000 visitors per variation.

Step 4: Run the Test

Split traffic evenly between Version A and Version B. Most A/B testing tools handle this automatically.

Critical rules during the test:

  • Do not peek at results early and make decisions
  • Do not stop the test as soon as one version looks like it is winning
  • Run the test for at least 7 full days to capture day-of-week effects
  • Do not change anything else on the page while the test is running

Step 5: Analyze Results

After reaching your required sample size:

  1. Check if the difference is statistically significant (p-value below 0.05)
  2. If significant, implement the winning version
  3. If not significant, the versions are effectively equal — keep whichever is simpler

A/B Testing Tools

Google Optimize (Sunset)

Google Optimize was the most popular free tool but was discontinued in 2023. Alternatives have filled the gap.

VWO

Visual editor for creating variations without code. Plans start at $199/month. Best for stores with significant traffic.

Convert

Privacy-focused testing tool with strong statistical rigor. Plans start at $99/month.

Manual Testing

For stores with limited budgets, you can run sequential tests: show Version A for one week, Version B the next week, and compare results. This is less rigorous but still better than guessing.

Common A/B Testing Mistakes

Stopping Tests Too Early

The most common mistake. You see Version B ahead after 100 visitors and declare it the winner. But with small samples, random chance creates apparent winners that would not hold up with more data. Always reach your calculated sample size.

Testing Too Many Things at Once

Multivariate testing (testing multiple elements simultaneously) requires exponentially more traffic. Unless you have tens of thousands of daily visitors, stick to single-variable A/B tests.

Ignoring Seasonal Effects

A test that runs only on weekdays will miss weekend behavior patterns. A test during a holiday sale does not represent normal performance. Run tests for at least one full week, preferably two.

Not Documenting Results

Keep a testing log with: what you tested, hypothesis, sample size, duration, results, and what you learned. This prevents retesting things you have already tested and builds institutional knowledge.

Testing Insignificant Elements

Do not spend two weeks testing whether a slightly different shade of blue performs better. Focus your limited traffic on tests that could meaningfully impact revenue.

A/B Testing Without Much Traffic

If your store gets fewer than 1,000 visitors per week, traditional A/B testing is difficult because reaching statistical significance takes too long.

Alternatives for low-traffic stores:

  • Sequential testing: Try Version A for two weeks, measure results, then try Version B for two weeks. Less rigorous but still data-informed.
  • Qualitative testing: Use heatmaps and session recordings to identify problems, then fix them directly.
  • Best practice implementation: Apply proven conversion optimization principles (covered in other guides) rather than testing from scratch.
  • Focus on traffic first: Once you have consistent traffic above 500-1000 sessions per week, start formal A/B testing.

Key Takeaways

  • A/B testing replaces guessing with evidence and is essential for optimization
  • Test one variable at a time and define your success metric before starting
  • Calculate required sample size before running any test to avoid false conclusions
  • Never stop a test early even if results look conclusive — wait for statistical significance
  • Prioritize high-impact elements like headlines, CTAs, and product images
  • Document everything to build a knowledge base of what works for your audience
  • Low-traffic stores should use qualitative methods until they have enough volume for proper testing

Ready to Put This Into Practice?

Launch your own fully automated dropshipping store and start applying these strategies today.