SEO Tests

SEO Utils allows you to run controlled experiments to measure the impact of your SEO changes using real Google Search Console data. The tool automatically tracks metrics like clicks, impressions, CTR, and rankings to help you make data-driven decisions.

SEO Tests: URL Switch Test.

You can test whether your SEO changes actually improve performance before rolling them out across your entire site. The tool supports three different test types, each designed for specific testing scenarios.

Test Types

SEO Utils supports three types of SEO tests:

  1. Time-based Test: Compare performance before and after implementing a change

  2. URL Switch Test: A/B test by alternating between two versions of the pages

  3. Split Test: Run simultaneous tests on different page groups

How to Create an SEO Test

To get started, head to the SEO Tests in the left sidebar. Then, click the Create Test button.

Access the SEO Tests tool in the left sidebar

The test creation process has three steps. We'll go through each step and explain the different fields for each test type.

Step 1: Configure Test

First, you need to set up the basic test information.

Configure your test settings

Common Fields (All Test Types):

  • Domain: Select the Google Search Console property you want to test

  • Test Type: Choose between Time-based, URL Switch, or Split Test

  • Test Name: Give your test a descriptive name

  • Description: Optional field to add more context about your test

  • Hypothesis: Clearly state what you expect to happen

Test Type Specific Fields:

For Time-based Tests:

  • Change Implemented Date: When the SEO change will be implemented. This date separates the "before" and "after" periods for comparison

For URL Switch Tests:

  • Switch Interval: How often to switch between versions (1-30 days)

  • Starting Version: Which version to show first (original or variant)

For Split Tests:

  • Traffic Split: Percentage of traffic for the variant group (10-90%)

Your hypothesis should be specific and measurable. For example: "Adding FAQ schema will increase CTR by 10%" instead of "Schema will help rankings".

Step 2: Select Pages

Next, you'll select which pages to include in your test.

Select and validate pages for testing

Page Selection Process:

  1. Add Filters: Use filters to find pages that match your criteria

  2. Date Range: Select the date range for baseline data validation

    • This ensures your selected pages have enough historical data

    • Recommended: Use at least 30 days of baseline data

  3. Find Pages: Click to search for matching pages in your GSC data

For Time-based Tests:

  • You can select or remove pages using checkboxes

For URL Switch Tests:

  • Configure URL pairs (original vs. alternate versions)

  • Example: /product (original) vs. /product-v2 (alternate)

  • The tool will automatically switch between these URLs based on your interval

For Split Tests:

  • Pages are automatically divided into Control and Variant groups

  • You can manually move pages between groups if needed

  • The tool ensures both groups have similar baseline performance

Step 3: Schedule Test

Finally, set when your test should run.

Schedule your test duration

Scheduling Fields:

  • Start Date: When to begin tracking test metrics

  • End Date: Optional - when to stop the test (leave empty for ongoing tests)

For Time-based Tests only:

  • Change Implemented Date: When you plan to implement (or have implemented) the SEO change

    • This separates the "before" and "after" periods

    • Must be between the start and end dates

When to Use Each Test Type

Time-based Test

Time-based Test

Use when:

  • Making site-wide changes (like updating all title tags)

  • Testing changes that can't be reversed easily

  • You want to compare "before" and "after" performance

  • Historical comparison is more important than simultaneous testing

Example scenarios:

  • Adding schema markup to all product pages

  • Changing the URL structure site-wide

  • Implementing Core Web Vitals improvements

URL Switch Test

URL Switch Test

Use when:

  • You have two versions of the same page

  • You can programmatically switch between versions

  • You want to minimize the impact of external factors

  • Testing significant page redesigns

Example scenarios:

  • Testing two different page layouts

  • Comparing long-form vs. short-form content

  • Testing different internal linking structures

Split Test

Use when:

  • You have many similar pages (like product or category pages)

  • You want to test changes on a subset before full rollout

  • You need results faster than sequential testing

  • Testing template-level changes

Example scenarios:

  • Testing new title tag templates on product pages

  • Adding FAQ sections to half of your blog posts

  • Testing different meta description formats

Understanding Statistical Terms

SEO Utils uses statistical analysis to ensure your test results are reliable. Here are the key terms:

1. Confidence Score / Statistical Significance

What it means: The probability that the observed difference is real and not due to random variation.

  • 95% confidence = Only 5% chance the results are due to randomness

  • 99% confidence = Only 1% chance the results are due to randomness

How to use it: Wait for at least 95% confidence before making decisions based on test results.

2. P-Value

What it means: The probability of seeing these results if there was actually no difference between versions.

  • p < 0.05: Statistically significant (less than 5% chance of random occurrence)

  • p < 0.01: Highly significant (less than 1% chance of random occurrence)

How to use it: A p-value below 0.05 indicates your results are statistically significant.

3. Z-Score

What it means: How many standard deviations the test results are from the mean. It measures the magnitude of difference.

  • |z| > 1.96: Significant at 95% confidence level

  • |z| > 2.58: Significant at 99% confidence level

How to use it: Larger absolute z-scores indicate stronger evidence of a real difference.

4. Uplift

What it means: The percentage change in performance between control and variant groups.

  • Positive uplift: Variant performs better than control

  • Negative uplift: Control performs better than variant

Calculation: (Variant Metric - Control Metric) / Control Metric Γ— 100

Interpreting Results Correctly

When to Trust Your Results

βœ… Reliable results when:

  • Statistical significance is 95% or higher

  • The test has run for at least 2-4 weeks

  • You have sufficient sample size (usually 1000+ clicks per group)

  • No major external events occurred during the test

⚠️ Be cautious when:

  • Significance is between 90-95%

  • Sample size is small (< 500 clicks)

  • Test duration is less than 2 weeks

  • Major algorithm updates occurred during testing

❌ Don't trust results when:

  • Significance is below 90%

  • Very small sample size (< 100 clicks)

  • Test ran for less than a week

  • Site had technical issues during the test

Monitoring Test Progress

SEO Utils provides real-time monitoring features:

  1. Daily Metrics Updates: See how metrics change day by day

  2. Trend Charts: Visualize performance trends over time

  3. Pause/Resume Capability: Exclude problematic periods from analysis

  4. Export Options: Download detailed results for further analysis

Pro Tip: Check your test results weekly, but avoid making decisions too early. Early results can be misleading due to small sample sizes.

Best Practices

  1. Define clear hypotheses: Be specific about what you expect to change and by how much

  2. Ensure sufficient sample size: Include enough pages and traffic for reliable results

  3. Run tests for adequate duration: At least 2-4 weeks, accounting for weekly patterns

  4. Monitor external factors: Note any algorithm updates or seasonal changes during your test

  5. Document your changes: Keep detailed records of exactly what was changed

Last updated