Skip to content

A/B Testing

Optimize your marketplace listing performance by comparing different content variants and measuring real user engagement.

Overview

A/B testing allows you to scientifically test different versions of your marketplace listing content to determine which performs better. By splitting traffic between variants and measuring key metrics, you can make data-driven decisions to improve your conversion rates.

Getting Started

Creating an A/B Test

  1. Navigate to GTM Campaigns > A/B Testing in your dashboard
  2. Click Create A/B Test
  3. Configure your test settings:
  4. Test Name: A descriptive name (e.g., "Title Optimization Test - March 2024")
  5. Hypothesis: What you expect to learn from this test
  6. Primary Metric: The main metric to determine the winner
  7. Select 2-4 content variants to compare
  8. Click Create A/B Test to start

Selecting Variants

You can test any content versions you've created for your marketplace listings. Select at least 2 variants (up to 4) to compare. Variants can include different:

  • Titles and taglines
  • Short descriptions
  • Feature highlights
  • Pricing presentation
  • Call-to-action messaging

Metrics

Click-Through Rate (CTR)

Formula: Clicks ÷ Impressions × 100

Measures how often viewers click on your listing when they see it in search results or category pages. A higher CTR indicates your listing is attracting more attention.

Conversion Rate

Formula: Subscriptions ÷ Page Views × 100

Measures how often viewers become paying customers after visiting your listing page. This is often the most important metric for revenue optimization.

Free Trials Started

Total number of trial signups attributed to each variant during the test period. Useful for products with a freemium or trial-based model.

Revenue

Total Monthly Recurring Revenue (MRR) generated by customers who converted during the test. This metric accounts for different pricing tiers and helps optimize for revenue rather than just volume.

Statistical Confidence

We use a two-tailed z-test to determine if differences between variants are statistically significant. Results are considered significant at the 95% confidence level.

What does 95% confidence mean?

When we say a result has 95% confidence, it means there's only a 5% probability that the observed difference occurred by random chance. This is the industry standard for making business decisions based on test results.

Confidence Indicators

Confidence Level Indicator Recommendation
Below 90% Low Continue testing
90-95% Moderate Consider extending test
95%+ High Safe to declare winner
99%+ Very High Strong statistical evidence

Best Practices

1. Test One Element at a Time

For clearer results, change only one element per test (title, description, or highlights). Testing multiple changes simultaneously makes it difficult to determine which change drove the improvement.

2. Run Tests Long Enough

Allow tests to run for at least 7-14 days to gather sufficient data. This ensures you capture:

  • Weekday vs. weekend traffic patterns
  • Enough conversions for statistical significance
  • Natural variation in user behavior

3. Wait for Statistical Confidence

Don't declare a winner until you've reached 95% statistical confidence. Early results can be misleading due to random variation.

4. Document Your Hypotheses

Keep records of what you tested and learned. This builds institutional knowledge and prevents repeating failed experiments.

5. Consider Sample Size

Ensure each variant receives enough traffic to generate meaningful data. Low-traffic listings may need longer test durations.

Interpreting Results

Reading the Results Dashboard

After your test has collected data, the results dashboard shows:

  • Variant Performance: Side-by-side comparison of each variant's metrics
  • Lift Percentage: How much better (or worse) each variant performed vs. baseline
  • Confidence Level: Statistical confidence that the difference is real
  • Sample Size: Number of impressions/views for each variant

Declaring a Winner

A variant can be declared the winner when:

  1. It outperforms other variants on the primary metric
  2. Statistical confidence reaches 95% or higher
  3. Sufficient sample size has been collected

No Clear Winner

Sometimes tests conclude without a clear winner. This happens when:

  • Differences between variants are too small to be statistically significant
  • All variants perform similarly
  • Sample size was insufficient

In these cases, consider:

  • Running a longer test
  • Testing more dramatically different variations
  • Focusing on a different element to optimize

Common Use Cases

Title Optimization

Test different title structures:

  • Feature-focused vs. benefit-focused
  • Including specific metrics or social proof
  • Different keyword placements

Description Testing

Compare description approaches:

  • Technical vs. business-focused language
  • Short vs. detailed descriptions
  • Different value proposition emphasis

Pricing Presentation

Test how you present pricing:

  • Highlighting savings vs. showing full price
  • Monthly vs. annual pricing emphasis
  • Including comparison to alternatives

FAQ

How long should I run a test?

We recommend a minimum of 7 days, ideally 14 days, to capture different traffic patterns and reach statistical significance.

Can I run multiple tests simultaneously?

Yes, but only if the tests don't overlap (different products or completely separate audiences). Running overlapping tests can contaminate results.

What if my listing doesn't get much traffic?

Low-traffic listings need longer test durations. Consider running tests for 30+ days or focusing on higher-traffic listings first.

Can I edit variants during a test?

No. Editing variants during a test invalidates the results. If you need to make changes, end the current test and start a new one.

How do I implement the winning variant?

Once a winner is declared, you can publish it directly from the results page. The winning content will replace your current live listing.