PageSpeed Matters
    PageSpeed Matters
    Tools & Data Sources · Glossary

    Field Data vs Lab Data · Definition & Explanation 2026

    Web performance measurement falls into two fundamentally different categories: field data (Real User Monitoring) and lab data (synthetic testing). Understanding the distinction is critical because Google uses field data — not lab scores — for ranking decisions, yet lab data is essential for debugging and optimization.

    Field data captures metrics from actual users in real-world conditions — diverse devices, network speeds, geographic locations, and browser configurations. The primary field data source is the Chrome User Experience Report (CrUX), which Google uses for the Page Experience ranking signal. Third-party RUM tools (SpeedCurve, DebugBear, Datadog) provide more granular field data.

    Lab data simulates a page load in a controlled environment — specific device, specific network, specific location. Lighthouse (which powers PageSpeed Insights' lab section) is the primary lab tool. Lab data is reproducible, fast for iteration, and provides detailed diagnostics.

    The common mistake: optimizing for Lighthouse scores (lab) while ignoring CrUX (field). A page can score 100 in Lighthouse but fail CWV in CrUX because real users on budget phones over 3G have a fundamentally different experience than the lab simulation.

    Updated 2026-02-28
    M
    By Matt Suffoletto

    TL;DR — Quick Summary

    Field data (CrUX, RUM) captures real users and is what Google uses for ranking. Lab data (Lighthouse) provides reproducible testing for debugging. Both are essential — use lab data for diagnosis and iteration, field data for measuring real impact.

    What is Field Data vs Lab Data?

    Field data collects metrics from actual users in real-world conditions — diverse devices, networks, and locations. The primary field source is CrUX (Chrome User Experience Report). RUM tools provide additional granularity.

    Lab data is collected in a controlled, simulated environment (Lighthouse, WebPageTest). Lab data is reproducible and provides detailed diagnostics; field data reflects reality but is aggregated.

    Key differences:

    • Field data = reality (but noisy, aggregated, delayed). Lab data = simulation (but reproducible, instant, detailed).
    • Field data is evaluated at p75 over 28 days. Lab data is a single snapshot.
    • Field captures all interactions (INP). Lab only captures loading (TBT as proxy).
    • Field reflects device/network diversity. Lab simulates one device/network.

    History & Evolution

    Key milestones:

    • 2005 — Early RUM tools emerge for measuring real-user performance.
    • 2012 — WebPageTest and Lighthouse establish lab testing as standard practice.
    • 2017 — CrUX launches, giving public access to Chrome field data.
    • 2020 — Core Web Vitals announced. Google explicitly states field data (CrUX) is used for rankings.
    • 2024 — INP replaces FID, highlighting a field-only metric (lab has no direct equivalent).
    • 2025–2026 — The field-vs-lab distinction is well understood. Best practice: diagnose with lab, measure with field.

    How Field Data vs Lab Data is Measured

    Field data tools:

    • CrUX (Google's public dataset, powers PSI and Search Console)
    • Web Vitals JS library (custom RUM)
    • SpeedCurve, DebugBear, Datadog, New Relic (commercial RUM)

    Lab data tools:

    • Lighthouse (Chrome DevTools, CLI, CI/CD)
    • WebPageTest (detailed waterfall analysis)
    • Chrome DevTools Performance panel

    Both:

    • PageSpeed Insights (shows CrUX field + Lighthouse lab in one view)

    Key rule: Field data (CrUX) determines Google rankings. Lab data (Lighthouse, WebPageTest) is for debugging and iteration.

    Common Causes of Poor Field Data vs Lab Data Scores

    Common causes of field-lab divergence:

    1. 1Device diversity — Lab tests one device; field includes budget phones with 2GB RAM.
    2. 2Network diversity — Lab simulates 4G; field includes 3G, congested networks.
    3. 3Geographic diversity — Lab tests from one location; field includes users far from servers.
    4. 4Browser extensions — Lab uses clean Chrome; real users have ad blockers, password managers.
    5. 5User behavior — Lab tests cold loads; field includes back-forward navigations, cached visits.
    6. 6Third-party scripts — Lab may not trigger all third-party loads; field captures full impact.

    Frequently Asked Questions

    Struggling with Field Data vs Lab Data?

    Request a free speed audit and we'll identify exactly what's holding your scores back.