Probeo
probeo

Core Web Vitals: CLS, INP, and LCP Risk Levels

How Probeo evaluates Core Web Vitals risk levels for CLS, INP, and LCP. Covers thresholds, measurement approaches, and what elevated risk signals mean for page experience.

Last updated 02/08/2026

Core Web Vitals are three metrics Google uses as page experience ranking signals: Cumulative Layout Shift (CLS) measures layout stability, Interaction to Next Paint (INP) measures responsiveness, and Largest Contentful Paint (LCP) measures loading speed. Each metric has defined thresholds that classify pages as good, needs improvement, or poor. Probeo evaluates structural risk factors for each metric based on page composition rather than synthetic test runs.

What's wrong

One or more Core Web Vitals metrics show elevated risk based on page structure analysis. This does not mean the page has failed a Lighthouse test. It means the page contains patterns known to produce poor CWV outcomes at scale: layout-shifting elements without reserved dimensions, long-blocking tasks in the render path, or late-loading primary content. These patterns may pass lab tests under ideal conditions while degrading consistently for real users on slower connections or constrained devices.

CLS: Cumulative Layout Shift

CLS quantifies how much visible content moves unexpectedly during page load and interaction. A CLS score below 0.1 is classified as good. Between 0.1 and 0.25 is needs improvement. Above 0.25 is poor. Layout shifts are caused by images without explicit dimensions, dynamically injected content above the fold, web fonts that swap with different metrics, or ad slots that resize after render. Each shift is scored by the fraction of the viewport affected multiplied by the distance moved. CLS accumulates across the entire page session, not just initial load.

INP: Interaction to Next Paint

INP measures the delay between a user interaction and the next visual update. It replaced First Input Delay (FID) as a Core Web Vital in March 2024. An INP below 200 milliseconds is good. Between 200ms and 500ms is needs improvement. Above 500ms is poor. Unlike FID, which only measured the first interaction, INP considers all interactions throughout the page session and reports the worst observed delay. Long INP values indicate the main thread is blocked by JavaScript execution, layout recalculations, or synchronous operations during user interaction.

LCP: Largest Contentful Paint

LCP measures when the largest visible content element finishes rendering. An LCP below 2.5 seconds is good. Between 2.5s and 4s is needs improvement. Above 4s is poor. The largest element is typically a hero image, heading block, or video poster frame. LCP is affected by server response time, render-blocking resources, resource load time for the LCP element itself, and client-side rendering delays. A page may report fast DOMContentLoaded while the content users actually see loads seconds later.

Threshold summary

  • CLS — Good: below 0.1 · Needs improvement: 0.1 to 0.25 · Poor: above 0.25
  • INP — Good: below 200ms · Needs improvement: 200ms to 500ms · Poor: above 500ms
  • LCP — Good: below 2.5s · Needs improvement: 2.5s to 4s · Poor: above 4s
  • Google uses the 75th percentile of field data to classify pages. A page passes CWV assessment when all three metrics are at or below the good threshold at the 75th percentile.

How Probeo evaluates risk versus lab and field measurement

Probeo does not run Lighthouse or collect real-user metrics. It identifies structural patterns in page markup, resource loading order, and DOM composition that correlate with poor CWV outcomes. This is a risk assessment, not a performance score. Lab tools like Lighthouse measure a single page load under controlled conditions. Field data from the Chrome User Experience Report (CrUX) reflects real user outcomes aggregated over 28 days. Both are valuable but neither reveals why a page is at risk before users encounter problems. Probeo operates upstream of both: it surfaces the structural causes that lab tests may miss and field data will eventually confirm.

Why it matters

Core Web Vitals are a confirmed Google ranking signal within the page experience system. They are one factor among many, and their impact should not be overstated. A page with poor CWV will not necessarily rank poorly if content relevance and authority are strong. But when competing pages are otherwise equivalent, CWV performance becomes a differentiator. More practically, pages that fail CWV thresholds deliver measurably worse user experience. Layout shifts erode trust. Slow interactions feel broken. Late-loading content frustrates users before they engage with it.

The correct change

Address the specific structural patterns flagged for each metric. For CLS: ensure all media elements have explicit width and height attributes, reserve space for dynamically loaded content, and avoid inserting content above existing visible elements. For INP: reduce main-thread blocking work during interactions, defer non-critical JavaScript, and avoid synchronous layout recalculations in event handlers. For LCP: prioritize loading of the largest content element, eliminate render-blocking resources from the critical path, and ensure the LCP resource begins loading early.

Scope

Core Web Vitals are evaluated per page. Google assesses CWV at the URL level using field data, and groups pages for assessment when individual URL data is insufficient. Templates that serve many pages amplify the impact of structural patterns. A CLS-causing layout pattern in a product template affects every product page. Fixing the template fixes the entire group.

How to verify

  • Confirmation that CWV risk factors have been resolved:
  • CLS risk elements have explicit dimensions or reserved layout space
  • INP risk factors show reduced main-thread blocking during interactions
  • LCP element loads within the critical rendering path without unnecessary delays
  • PageSpeed Insights or CrUX data shows metrics within good thresholds at the 75th percentile
  • Structural risk indicators in Probeo are cleared or reduced in severity

Takeaway

  • Core Web Vitals measure layout stability, responsiveness, and loading speed
  • Each metric has defined good, needs-improvement, and poor thresholds
  • Google uses the 75th percentile of real user data for ranking assessment
  • CWV is one ranking signal among many — it does not override content relevance
  • Probeo identifies structural risk patterns that correlate with poor CWV outcomes
  • Fixing root causes in templates scales improvements across all pages using that template

FAQ

What is the difference between lab data and field data for Core Web Vitals?
Lab data comes from synthetic tests like Lighthouse running a single page load under controlled conditions. Field data comes from real users via the Chrome User Experience Report, aggregated over 28 days. Google uses field data for ranking decisions. Lab data is useful for debugging but does not reflect the variance of real user conditions: device capability, network speed, and interaction patterns.
Do Core Web Vitals thresholds differ between mobile and desktop?
The threshold values are the same for both mobile and desktop. However, Google evaluates mobile and desktop separately using distinct field data sets. A page may pass CWV on desktop but fail on mobile due to differences in rendering performance, network conditions, and interaction patterns. Mobile field data typically shows worse INP and LCP values than desktop.
How much do Core Web Vitals actually affect search rankings?
CWV is one signal within the broader page experience system, which itself is one factor among hundreds. Google has stated that content relevance remains the primary ranking factor. CWV acts as a tiebreaker when competing pages have similar content quality and authority. Overstating CWV's ranking impact leads to misallocated optimization effort.
Why might a page pass Lighthouse but fail in field data?
Lighthouse tests a single load on a controlled environment with consistent CPU and network settings. Field data captures the full distribution of real user experiences across varied devices, connections, and interaction timing. A page that scores well under lab conditions may perform poorly for users on mid-range Android devices over cellular connections, which often represent the 75th percentile.
Why did Google replace FID with INP?
First Input Delay only measured the delay of the first user interaction, ignoring all subsequent interactions. A page could report excellent FID while remaining unresponsive throughout the rest of the session. INP measures responsiveness across all interactions and reports the worst case, providing a more complete picture of how the page actually responds to user input.
Does Probeo replace Lighthouse or PageSpeed Insights?
No. Probeo identifies structural risk factors in page composition that tend to produce poor CWV outcomes. It operates before measurement tools run. Lighthouse and PageSpeed Insights confirm whether those outcomes materialize. Probeo surfaces causes. Measurement tools confirm effects. They serve different purposes in the same workflow.