PageSpeed Matters
    Speed Audit
    Let's Talk
    PageSpeed Matters
    Book a Call
    Tools & Data Sources · Glossary

    Field Data vs Lab Data · Definition & Explanation 2026

    Web performance measurement falls into two fundamentally different categories: field data (Real User Monitoring) and lab data (synthetic testing). Understanding the distinction is critical because Google uses field data — not lab scores — for ranking decisions, yet lab data is essential for debugging and optimization.

    Field data captures metrics from actual users in real-world conditions — diverse devices, network speeds, geographic locations, and browser configurations. The primary field data source is the Chrome User Experience Report (CrUX), which Google uses for the Page Experience ranking signal. Third-party RUM tools (SpeedCurve, DebugBear, Datadog) provide more granular field data.

    Lab data simulates a page load in a controlled environment — specific device, specific network, specific location. Lighthouse (which powers PageSpeed Insights' lab section) is the primary lab tool. Lab data is reproducible, fast for iteration, and provides detailed diagnostics.

    The common mistake: optimizing for Lighthouse scores (lab) while ignoring CrUX (field). A page can score 100 in Lighthouse but fail CWV in CrUX because real users on budget phones over 3G have a fundamentally different experience than the lab simulation.

    Updated 2026-02-28
    M
    By Matt Suffoletto

    TL;DR — Quick Summary

    Field data (CrUX, RUM) captures real users and is what Google uses for ranking. Lab data (Lighthouse) provides reproducible testing for debugging. Both are essential — use lab data for diagnosis and iteration, field data for measuring real impact.

    What is Field Data vs Lab Data?

    Field data collects metrics from actual users in real-world conditions — diverse devices, networks, and locations. The primary field source is CrUX (Chrome User Experience Report). RUM tools provide additional granularity.

    Lab data is collected in a controlled, simulated environment (Lighthouse, WebPageTest). Lab data is reproducible and provides detailed diagnostics; field data reflects reality but is aggregated.

    Key differences:

    • Field data = reality (but noisy, aggregated, delayed). Lab data = simulation (but reproducible, instant, detailed).
    • Field data is evaluated at p75 over 28 days. Lab data is a single snapshot.
    • Field captures all interactions (INP). Lab only captures loading (TBT as proxy).
    • Field reflects device/network diversity. Lab simulates one device/network.

    History & Evolution

    Key milestones:

    • 2005 — Early RUM tools emerge for measuring real-user performance.
    • 2012 — WebPageTest and Lighthouse establish lab testing as standard practice.
    • 2017 — CrUX launches, giving public access to Chrome field data.
    • 2020 — Core Web Vitals announced. Google explicitly states field data (CrUX) is used for rankings.
    • 2024 — INP replaces FID, highlighting a field-only metric (lab has no direct equivalent).
    • 2025–2026 — The field-vs-lab distinction is well understood. Best practice: diagnose with lab, measure with field.

    How Field Data vs Lab Data is Measured

    Field data tools:

    • CrUX (Google's public dataset, powers PSI and Search Console)
    • Web Vitals JS library (custom RUM)
    • SpeedCurve, DebugBear, Datadog, New Relic (commercial RUM)

    Lab data tools:

    • Lighthouse (Chrome DevTools, CLI, CI/CD)
    • WebPageTest (detailed waterfall analysis)
    • Chrome DevTools Performance panel

    Both:

    • PageSpeed Insights (shows CrUX field + Lighthouse lab in one view)

    Key rule: Field data (CrUX) determines Google rankings. Lab data (Lighthouse, WebPageTest) is for debugging and iteration.

    Common Causes of Poor Field Data vs Lab Data Scores

    Common causes of field-lab divergence:

    1. 1Device diversity — Lab tests one device; field includes budget phones with 2GB RAM.
    2. 2Network diversity — Lab simulates 4G; field includes 3G, congested networks.
    3. 3Geographic diversity — Lab tests from one location; field includes users far from servers.
    4. 4Browser extensions — Lab uses clean Chrome; real users have ad blockers, password managers.
    5. 5User behavior — Lab tests cold loads; field includes back-forward navigations, cached visits.
    6. 6Third-party scripts — Lab may not trigger all third-party loads; field captures full impact.

    Frequently Asked Questions

    Field data (CrUX). Google explicitly uses CrUX field data for the Page Experience ranking signal. Lighthouse lab scores do not directly affect rankings.

    Lighthouse simulates one device on one network. CrUX aggregates millions of real users on diverse devices and networks. The p75 of real users is often worse than the lab simulation.

    Yes — very common. If many of your real users have slow devices or connections, field data will be worse than lab data. This means you need to optimize for the long tail of slower experiences.

    Test on budget devices (Moto G Power, Samsung A-series). Throttle to 3G in DevTools. Optimize for the slowest 25% of your user base, not just the median.

    WebPageTest is lab data — it runs controlled tests from specific locations and devices. It does not capture real-user data.

    RUM tools are field data — they capture metrics from real users. They provide more granular data than CrUX (per-page, per-device, per-region breakdowns) but aren't used by Google for ranking.

    No — Lighthouse is essential for debugging. Use Lighthouse to identify problems and verify fixes quickly. Then wait for CrUX field data to confirm the improvement is reflected in real-user experience.

    CrUX uses a 28-day rolling window. Improvements begin showing immediately as new fast data enters the window, but full improvement takes 28 days as old slow data rolls out.

    Struggling with Field Data vs Lab Data?

    Request a free speed audit and we'll identify exactly what's holding your scores back.

    Request An Audit