PageSpeed Matters
    PageSpeed Matters
    Comparison Guide

    PageSpeed Insights vs GTmetrix vs WebPageTest vs Lighthouse: Best Speed Testing Tool 2026

    Updated March 1, 2026
    22 min Read
    M
    Matt Suffoletto
    P
    PageSpeed Matters Team

    You run your URL through PageSpeed Insights and get a score of 62. Then you test on GTmetrix and see an A grade with 94%. WebPageTest shows a Largest Contentful Paint of 1.8s — while your CrUX data says real users experience 3.2s. Which number do you trust? Which tool should drive your optimization decisions?

    This confusion isn't a bug — it's by design. Each tool measures different things, from different locations, under different conditions, using different scoring algorithms. Understanding these differences is the gap between wasting weeks chasing the wrong metric and making targeted improvements that actually move revenue.

    We've used all four tools across 500+ enterprise optimization projects. This comparison reveals exactly when to use each tool, what their scores actually mean, and which combination gives you the complete performance picture.

    Why this matters in 2026: Google uses CrUX field data — not Lighthouse lab scores — for ranking signals. Yet most teams still optimize exclusively for Lighthouse scores. This disconnect means you could have a perfect 100 Lighthouse score and still fail Core Web Vitals in Google Search Console. Understanding which tool maps to ranking impact is worth more than any single optimization.

    Need expert help?

    TL;DR — Quick Summary

    Quick verdict: No single tool is "best" — you need at least two. PageSpeed Insights (PSI) is the only tool that combines lab data (Lighthouse) with real-user field data (CrUX) and directly maps to Google's ranking signals. Use PSI as your primary tool, WebPageTest for deep diagnostics, and GTmetrix for client-friendly reporting.

    Tool selection by use case:

    • SEO / rankings impact: PageSpeed Insights (shows CrUX data Google uses for rankings)
    • Deep technical debugging: WebPageTest (waterfall, filmstrip, connection throttling)
    • Client reporting: GTmetrix (clean UI, historical tracking, A-F grades)
    • CI/CD automation: Lighthouse CLI (runs in pipelines, budget enforcement)
    • Quick health check: PageSpeed Insights (fastest, no setup)

    Critical distinction: Lab data (Lighthouse, WebPageTest, GTmetrix) measures potential performance under controlled conditions. Field data (CrUX via PSI) measures actual user experience. Only field data affects Google rankings.

    Key Takeaways

    • PageSpeed Insights is the only tool showing CrUX field data — the actual data Google uses for page experience ranking signals. Always check PSI's 'Discover what your real users are experiencing' section first.
    • Lighthouse scores vary 5-15 points between runs due to CPU/network variability. Never chase a specific number — focus on consistent improvement trends and passing Core Web Vitals thresholds.
    • GTmetrix tests from a single location (Vancouver by default) with a desktop viewport. Its scores don't reflect mobile user experience or global performance without manual configuration changes.
    • WebPageTest is the gold standard for deep diagnostics — waterfall analysis, filmstrip comparison, custom connection throttling, and multi-step scripted tests that no other tool matches.
    • A site can score 95+ on Lighthouse but fail CrUX Core Web Vitals. Lab scores test under ideal conditions; field data captures real device diversity, network conditions, and user behavior patterns.
    • For Google rankings, only CrUX field data matters. Lighthouse/lab scores are diagnostic tools for finding what to fix — not a direct reflection of ranking impact.

    Quick Comparison Table: Speed Testing Tools 2026

    Here's the high-level comparison of all four tools:

    FeaturePageSpeed Insights (PSI)GTmetrixWebPageTestLighthouse (CLI/DevTools)
    Data TypeLab + Field (CrUX)Lab onlyLab onlyLab only
    Scoring0-100 (Lighthouse) + CWV pass/failA-F grade + 0-100%No single score (metric-based)0-100
    Mobile Testing✅ Default view✅ (must configure)✅ (device emulation)✅ Default
    Desktop Testing✅ Toggle available✅ Default view✅ Default✅ Available
    Test LocationGoogle servers30+ locations (Vancouver default)40+ locations globallyLocal machine or CI
    Connection ThrottleFixed (Moto G Power on 4G)ConfigurableFully configurableConfigurable
    Waterfall Analysis❌ Basic treemap only✅ Good✅ Best-in-class❌ Basic
    Filmstrip View✅ Best-in-class
    Historical Tracking❌ (CrUX is 28-day rolling)✅ (paid plans)✅ (free)❌ (manual)
    API Access✅ Free✅ (paid plans)✅ Free✅ Free (npm)
    CI/CD IntegrationVia APILimitedVia API✅ Native (CLI)
    CrUX Field Data✅ Built-in❌ (Chrome UX extension)
    CostFreeFreemium ($14.95–$49.95/mo)Free (sponsored)Free
    Maps to Rankings✅ (CrUX section)
    Best ForSEO impact, quick checksClient reporting, monitoringDeep debugging, comparisonsCI/CD, development

    Key insight: PageSpeed Insights is the only tool that shows what Google actually sees (CrUX data). The others are diagnostic tools that help you find and fix issues — but their scores don't directly reflect ranking impact.

    Lab Data vs Field Data: The Most Misunderstood Concept

    This distinction is the single most important thing to understand about speed testing. Getting it wrong means optimizing for the wrong target.

    Lab Data (Synthetic Testing):

    • Controlled environment: specific device, network, location, browser
    • Reproducible (mostly) — same conditions each test
    • Shows potential performance under those specific conditions
    • Tools: Lighthouse, GTmetrix, WebPageTest
    • Used for: Diagnosing issues, testing changes, comparing before/after
    • Does NOT directly affect Google rankings

    Field Data (Real User Monitoring / RUM):

    • Actual measurements from real users visiting your site
    • Aggregated across all devices, networks, locations, browsers
    • Shows actual user experience across the full distribution
    • Sources: CrUX (Chrome UX Report), custom RUM (SpeedCurve, mPulse)
    • Used for: Understanding true user experience, tracking ranking signals
    • ✅ Directly affects Google rankings (via page experience signals)

    Why They Diverge — Real Example:

    A client's homepage scored 92 on Lighthouse but had a 3.8s LCP in CrUX. Why?

    • Lighthouse tested from a US server on simulated Moto G Power with stable 4G
    • Real users included: visitors on 3G in India (40% of traffic), older Android devices with 2GB RAM, users with ad blockers disabled (loading all 18 tracking scripts)
    • The lab test couldn't simulate the long tail of real-world conditions

    The P75 Problem: CrUX reports the 75th percentile — meaning 75% of real user experiences are at or better than the reported value. That bottom 25% (often users on slow connections or old devices) can drag your CWV assessment to 'Needs Improvement' even if most users have a fast experience.

    Practical Rule:

    1. 1Use field data (CrUX via PSI) to know where you stand with Google
    2. 2Use lab data (Lighthouse, WebPageTest) to diagnose why and what to fix
    3. 3After fixing, verify improvement in lab → then wait 28 days for field data to update

    For a deeper dive, see our CrUX field data guide and Lighthouse guide.

    PageSpeed Insights (PSI): The SEO Essential

    PageSpeed Insights is Google's own tool — and the only one that combines Lighthouse lab analysis with CrUX field data in a single view.

    What PSI Actually Shows:

    Section 1 — 'Discover what your real users are experiencing' (CrUX):

    • This is the data Google uses for page experience ranking signals
    • Shows LCP, INP, CLS, FCP, TTFB at the 75th percentile
    • Green/amber/red indicators match Google Search Console's Core Web Vitals report
    • Data is a 28-day rolling average — changes take ~28 days to reflect
    • Available at origin level (entire domain) and URL level (if enough traffic)
    • If this section says 'Not enough real-world speed data' — your page doesn't have enough Chrome traffic for URL-level CrUX. Check origin-level data instead.

    Section 2 — Performance Score (Lighthouse):

    • Standard Lighthouse 0-100 score run from Google's servers
    • Simulates Moto G Power on throttled 4G connection (mobile view)
    • Desktop view uses unthrottled connection with desktop viewport
    • Score is a weighted composite: LCP (25%), TBT (30%), CLS (25%), FCP (10%), Speed Index (10%)
    • See our PageSpeed Insights guide for detailed scoring breakdown

    Section 3 — Opportunities & Diagnostics:

    • Actionable recommendations sorted by estimated impact
    • 'Opportunities' show potential time savings (e.g., 'Serve images in next-gen formats — estimated savings 2.4s')
    • 'Diagnostics' show additional information (DOM size, main-thread work, etc.)
    • Passed audits show what you're already doing well

    PSI Strengths:

    • Only tool connecting lab analysis to ranking-relevant field data
    • Free, no account needed, instant results
    • Consistent test environment (Google's infrastructure)
    • API available for programmatic access (500 requests/day free)

    PSI Limitations:

    • No waterfall diagram — can't see request-level loading sequence
    • No filmstrip — can't see visual loading progression
    • No historical tracking — can't compare tests over time
    • Single test location (Google servers) — can't test from specific regions
    • Lab scores vary 5-15 points between runs (CPU/network variation on Google's shared infrastructure)
    • CrUX data requires sufficient Chrome traffic (~1,000+ page views/month per URL)

    When to Use PSI:

    • First check for any page — see if CrUX data passes CWV thresholds
    • After optimizations — verify Lighthouse score improved
    • SEO discussions — this is the data stakeholders and Google care about
    • Quick competitive analysis — compare your CrUX vs competitors

    Pro Tip: Always check the mobile view first. Google uses mobile-first indexing, and mobile CrUX data is what affects rankings. Desktop scores are nice-to-have but rarely impact SEO.

    GTmetrix: The Client-Friendly Reporter

    GTmetrix is the most visually polished speed testing tool, making it popular for client reporting and monitoring. It runs Lighthouse under the hood but adds its own grading, waterfall, and monitoring features.

    What GTmetrix Actually Measures:

    • Runs Lighthouse in a real Chrome browser (not simulated)
    • Default test: Desktop viewport, Vancouver (Canada) location, unthrottled connection
    • Generates an A-F grade based on GTmetrix Structure score + Lighthouse Performance score
    • Shows Web Vitals (LCP, TBT, CLS) plus legacy metrics (Fully Loaded Time, Total Page Size, Requests)

    GTmetrix Grade vs Lighthouse Score — Why They Differ:

    GTmetrix's letter grade combines two components:

    1. 1GTmetrix Structure (50%) — audits for best practices (image optimization, caching headers, minification)
    2. 2Performance (50%) — Lighthouse Performance score

    This means a site with excellent structure but mediocre speed can still get a B+ on GTmetrix while scoring 55 on Lighthouse. The grade can be misleading if taken at face value.

    GTmetrix Strengths:

    • Clean, professional reports — excellent for client presentations
    • Waterfall diagram with resource-level timing breakdown
    • Video filmstrip showing visual loading progression
    • Historical monitoring (paid plans) with alerting
    • 30+ test locations (paid plans)
    • Real browser testing (not simulated throttling like Lighthouse)

    GTmetrix Limitations:

    • ❌ No CrUX field data — scores don't reflect real user experience or ranking impact
    • Default desktop test doesn't match Google's mobile-first evaluation
    • Free tier limited to Vancouver location — not useful for global sites
    • Grades can give false confidence (A grade ≠ good CrUX performance)
    • Paid plans required for mobile testing, monitoring, and non-Vancouver locations
    • Tests run on GTmetrix's infrastructure — results may differ significantly from PSI

    GTmetrix vs PSI Score Differences — Common Scenarios:

    ScenarioPSI Mobile ScoreGTmetrix GradeWhy They Differ
    Fast CDN, heavy JS55A (92%)GTmetrix tests desktop (no CPU throttle); PSI throttles to Moto G
    Image-heavy, good structure70A (95%)GTmetrix Structure score inflates grade
    Server in US, global audience85B (78%)GTmetrix from Vancouver; PSI from Google servers (closer to CDN edges)
    Lightweight, poor caching90C (72%)GTmetrix penalizes missing cache headers heavily in Structure

    When to Use GTmetrix:

    • Client reporting — the cleanest, most presentable reports
    • Before/after comparisons — visual filmstrip is compelling for stakeholders
    • Monitoring over time — paid plans track performance trends
    • Waterfall analysis — better than PSI (which has none)

    When NOT to Use GTmetrix:

    • As your primary SEO performance indicator — it doesn't show CrUX data
    • For mobile performance assessment — default is desktop
    • As a replacement for PSI — always cross-reference with PSI for ranking context

    For detailed optimization guidance, see our GTmetrix guide.

    Need help with speed optimization?

    Our team specializes in performance optimization. Request an audit and see how much faster your site could be.

    WebPageTest: The Deep Diagnostics Powerhouse

    WebPageTest is the most technically powerful speed testing tool available. Created by Patrick Meenan (formerly of Google), it's the gold standard for performance engineers who need to understand exactly what's happening during page load.

    What Makes WebPageTest Unique:

    • Tests from 40+ real locations worldwide using real browsers on real devices
    • Fully configurable: connection speed, latency, packet loss, CPU throttle
    • Multi-step scripted testing (login → navigate → interact → measure)
    • Comparative testing (test two URLs side-by-side with identical conditions)
    • First View vs Repeat View separation (cold cache vs warm cache)
    • Request-level detail: DNS, connect, TLS, TTFB, download for every single resource

    WebPageTest's Best Features:

    Waterfall Diagram (Best-in-Class):

    • Every HTTP request visualized with DNS/connect/TLS/TTFB/download breakdown
    • Color-coded by resource type (HTML, CSS, JS, images, fonts, third-party)
    • Vertical lines marking key timing milestones (Start Render, LCP, DOM Complete)
    • Click any request to see full headers, response body, and timing
    • Third-party requests highlighted — instantly see which external scripts delay loading

    Filmstrip View:

    • Visual screenshots captured every 100ms during page load
    • Side-by-side comparison of two URLs loading simultaneously
    • Instantly identifies when content becomes visible vs when metrics fire
    • Excellent for spotting CLS — you can see layout shifts frame by frame

    Connection Profiles:

    • Simulate exact network conditions: 3G Slow, 3G Fast, 4G, Cable, FIOS, Custom
    • Set specific bandwidth, latency, and packet loss values
    • Match real-world conditions your users experience
    • Test how your site degrades on poor connections

    Scripted Multi-Step Tests: ``` navigate https://example.com setValue id=email test@example.com setValue id=password test123 click id=login-button waitForComplete navigate https://example.com/dashboard ``` Test authenticated pages, checkout flows, and multi-page user journeys.

    WebPageTest Strengths:

    • Deepest technical analysis of any tool — unmatched waterfall and filmstrip
    • 40+ global test locations with real devices
    • Free and open-source (self-hosted option available)
    • API for automation (free, rate-limited)
    • No scoring bias — shows raw metrics without opinionated grading
    • Security headers analysis, HTTP/2 priority visualization, font loading analysis

    WebPageTest Limitations:

    • ❌ No CrUX field data — lab-only measurements
    • Steeper learning curve — the UI is functional, not polished
    • Results require expertise to interpret — no simple pass/fail
    • Queue times can be long during peak hours (5-15 minutes)
    • Reports aren't client-friendly without explanation
    • No built-in historical tracking (must use API + external storage)

    When to Use WebPageTest:

    • Deep debugging — when PSI says 'Reduce unused JavaScript' but you need to know exactly which scripts
    • Third-party script analysis — identify which external scripts block rendering
    • Before/after optimization comparison — identical conditions, side-by-side filmstrip
    • Regional performance testing — test from user locations, not just US servers
    • Font loading analysis — detailed font loading waterfall
    • HTTP/2 and HTTP/3 analysis — connection-level diagnostics

    Pro Tip: Use the 'Simple Testing' view for quick tests and 'Advanced Testing' when you need specific connection profiles, scripted tests, or custom headers. Always run 3 tests and use the median result — single runs have high variance.

    Lighthouse (CLI / DevTools): The Developer's Daily Driver

    Lighthouse is the open-source engine that powers PageSpeed Insights and GTmetrix. Running it directly via Chrome DevTools or the CLI gives you the most control over test conditions.

    Three Ways to Run Lighthouse:

    1. Chrome DevTools (F12 → Lighthouse tab):

    • Runs on your local machine with your CPU and network
    • Quick and convenient — no external tool needed
    • Results vary significantly based on your machine's specs and load
    • Best for: Quick development checks, not benchmarking

    2. Lighthouse CLI (npm): ```bash npm install -g lighthouse lighthouse https://example.com --output html --output-path ./report.html ```

    • Runs with configurable throttling (simulated or applied)
    • Consistent conditions when run in CI/CD pipelines
    • JSON output for programmatic analysis and budget enforcement
    • Best for: CI/CD integration, automated regression testing

    3. Lighthouse CI (LHCI): ```bash npm install -g @lhci/cli lhci autorun --collect.url=https://example.com --assert.preset=lighthouse:recommended ```

    • Purpose-built for CI/CD pipelines
    • Historical comparison against baselines
    • Performance budgets with pass/fail assertions
    • GitHub status checks integration
    • Best for: Automated performance regression prevention

    Lighthouse Scoring in 2026:

    MetricWeightGood Threshold
    Total Blocking Time (TBT)30%< 200ms
    Largest Contentful Paint (LCP)25%≤ 2.5s
    Cumulative Layout Shift (CLS)25%≤ 0.1
    First Contentful Paint (FCP)10%≤ 1.8s
    Speed Index10%≤ 3.4s

    Note: Lighthouse uses TBT as a proxy for INP in lab testing. Field INP and lab TBT don't always correlate perfectly — a page can have good TBT but poor field INP if interactions trigger expensive event handlers.

    Score Variability — The Elephant in the Room: Lighthouse scores vary 5-15 points between consecutive runs, even under identical conditions. Reasons:

    • CPU scheduling differences on the test machine
    • Network timing variations (even with throttling)
    • Third-party script timing (ad networks, chat widgets load unpredictably)
    • Background processes consuming resources

    Mitigation: Run 3-5 tests and use the median. In CI/CD, use Lighthouse CI's median assertion mode. Never make optimization decisions based on a single Lighthouse run.

    Lighthouse Strengths:

    • Free, open-source, runs anywhere (CLI, DevTools, CI/CD)
    • Detailed audit explanations with 'Learn more' links
    • Performance budgets for automated regression detection
    • Accessibility, SEO, and Best Practices audits alongside Performance
    • JSON output for custom dashboards and analysis
    • Same engine as PSI — consistent scoring methodology

    Lighthouse Limitations:

    • ❌ No CrUX field data (lab-only)
    • Score variability makes absolute numbers unreliable
    • Simulated throttling can under-represent real-world slowness
    • DevTools runs are influenced by extensions, other tabs, machine load
    • No waterfall diagram (use WebPageTest for that)
    • No visual filmstrip comparison

    When to Use Lighthouse:

    • Development workflow — quick checks during coding
    • CI/CD pipelines — automated regression prevention
    • Performance budgets — enforce JS size, image count, LCP thresholds
    • Comprehensive audits — Performance + Accessibility + SEO in one run

    For detailed Lighthouse optimization strategies, see our Lighthouse guide.

    Scoring Differences Explained: Why Your Numbers Don't Match

    The most common frustration: 'PSI says 62, GTmetrix says A, and my client wants to know which is right.' Both are right — they're measuring different things.

    Why Scores Differ Between Tools:

    1. Device & Throttling:

    • PSI Mobile: Simulates Moto G Power (mid-range) on throttled 4G (1.6Mbps down, 150ms RTT)
    • PSI Desktop: No CPU throttle, no network throttle
    • GTmetrix Default: Desktop viewport, unthrottled connection, real Chrome browser
    • WebPageTest: Fully configurable (you choose device, connection, location)
    • Lighthouse DevTools: Your machine's CPU (usually faster than Moto G simulation)

    2. Test Location:

    • PSI: Google's distributed infrastructure (generally close to CDN edges)
    • GTmetrix Free: Vancouver, Canada (far from many target audiences)
    • WebPageTest: 40+ locations globally (you choose)
    • Lighthouse CLI: Your machine's location

    3. Scoring Algorithm:

    • PSI/Lighthouse: Weighted composite (TBT 30%, LCP 25%, CLS 25%, FCP 10%, SI 10%)
    • GTmetrix: 50% GTmetrix Structure + 50% Lighthouse Performance
    • WebPageTest: No composite score — individual metrics only

    Common Score Gaps and What They Mean:

    Gap: PSI Mobile low, GTmetrix high Meaning: Your site is fast on desktop but struggles on mobile CPU/network throttling. Focus on JavaScript reduction (main-thread work) and image optimization. This is the most common gap.

    Gap: PSI high, GTmetrix low Meaning: Poor caching headers or missing optimizations that GTmetrix Structure penalizes. Add proper Cache-Control headers, enable text compression, implement resource hints.

    Gap: Lab scores high, CrUX failing Meaning: Real users on diverse devices/networks have a worse experience than lab simulations. Look at your traffic demographics — if 30%+ comes from emerging markets on 3G, lab tests with 4G throttling won't capture their experience. Implement adaptive loading or lighter experiences for slow connections.

    Gap: CrUX passing, Lighthouse low Meaning: Your real users are fine — don't panic about the Lighthouse score. This often happens when your audience is primarily on fast devices/connections (e.g., B2B SaaS with corporate users on fiber). Focus on maintaining CrUX performance, not chasing Lighthouse perfection.

    The Only Number That Matters for SEO: CrUX data in PageSpeed Insights. If the 'Discover what your real users are experiencing' section shows all green Core Web Vitals — your site passes Google's page experience assessment regardless of what any lab score says.

    Practical Workflow:

    1. 1Check PSI CrUX data → passing? Great for SEO. Not passing? Continue.
    2. 2Run PSI Lighthouse → identify top 3 opportunities with highest estimated savings
    3. 3Use WebPageTest → deep-dive into waterfall for the specific bottlenecks
    4. 4Fix → verify in Lighthouse CLI (faster iteration than PSI)
    5. 5Deploy → wait 28 days → verify CrUX data improved in PSI

    Repeat until CrUX passes. Then shift to monitoring.

    Building Your Testing Workflow: Which Tools When

    Here's the practical testing workflow we use across 500+ client projects.

    Phase 1: Initial Assessment (Day 1)

    ActionToolWhy
    Check CrUX statusPageSpeed InsightsSee if real users pass CWV — this is what Google sees
    Get mobile baselinePSI (mobile tab)Lighthouse score + top opportunities
    Get desktop baselinePSI (desktop tab)Usually higher; confirms mobile is the priority
    Deep waterfall analysisWebPageTest (mobile, 4G)Identify exact bottleneck requests
    Third-party script auditWebPageTestColor-coded third-party highlighting
    Generate client reportGTmetrixClean visuals for stakeholder presentation

    Phase 2: During Optimization (Weeks 1-4)

    ActionToolWhy
    Before/after each changeLighthouse CLI (3-run median)Fast iteration, consistent conditions
    Visual regression checkWebPageTest filmstripConfirm CLS hasn't increased
    CI/CD gateLighthouse CIBlock deploys that regress performance
    Weekly progress reportGTmetrix (monitoring)Show trend lines to stakeholders

    Phase 3: Post-Optimization Monitoring (Ongoing)

    ActionToolFrequency
    CrUX field data verificationPageSpeed InsightsMonthly (28-day rolling average)
    Regression monitoringGTmetrix or Lighthouse CIWeekly automated
    Deep diagnostic (if regression)WebPageTestAs needed
    Competitive benchmarkingPSI (compare CrUX)Quarterly

    CI/CD Integration Example (Lighthouse CI): ```json { "ci": { "collect": { "url": ["https://example.com/", "https://example.com/product/"], "numberOfRuns": 5 }, "assert": { "assertions": { "categories:performance": ["error", { "minScore": 0.7 }], "largest-contentful-paint": ["error", { "maxNumericValue": 3000 }], "cumulative-layout-shift": ["error", { "maxNumericValue": 0.1 }], "total-blocking-time": ["warn", { "maxNumericValue": 300 }] } } } } ```

    Budget Alerts (GTmetrix Pro): Set alerts for: LCP > 3.0s, CLS > 0.15, page weight > 3MB, third-party requests > 15. Get email notifications when regressions occur before users complain.

    Tool Cost Summary:

    ToolFree TierPaid PlansBest Value
    PageSpeed InsightsUnlimitedN/A (free)Always use — it's free and essential
    GTmetrix3 tests/day, desktop only$14.95–$49.95/moWorth it for agencies (monitoring + client reports)
    WebPageTestUnlimited (queued)Catchpoint RUM ($$$)Free tier is sufficient for most teams
    Lighthouse CLIUnlimited (local)N/A (free/open-source)Always use in CI/CD — zero cost

    The Minimum Viable Testing Stack:

    • PageSpeed Insights (CrUX + quick Lighthouse) — always
    • Lighthouse CI in your deployment pipeline — always
    • WebPageTest for deep dives — when diagnosing specific issues
    • GTmetrix for monitoring/reporting — if you need client-facing reports

    Thresholds & Benchmarks

    MetricGoodNeeds ImprovementPoor
    LCP (Largest Contentful Paint)≤ 2.5s2.5s – 4.0s> 4.0s
    INP (Interaction to Next Paint)≤ 200ms200ms – 500ms> 500ms
    CLS (Cumulative Layout Shift)≤ 0.10.1 – 0.25> 0.25
    TTFB (Time to First Byte)< 300ms300–800ms> 800ms
    FCP (First Contentful Paint)≤ 1.8s1.8s – 3.0s> 3.0s
    Speed Index≤ 3.4s3.4s – 5.8s> 5.8s
    Total Blocking Time (TBT)< 200ms200–600ms> 600ms
    Lighthouse Performance Score90+50–89Below 50

    Need help with speed optimization?

    Our team specializes in performance optimization. Request an audit and see exactly how much faster your site could be.

    Frequently Asked Questions

    Ready to Speed Up Your Site?

    Stop reading guides and start seeing results. Request an audit from our team today.

    Speed performance gauge showing optimized score