Core Web Vitals: CLS, INP, and LCP Risk Levels
How Probeo evaluates Core Web Vitals risk levels for CLS, INP, and LCP. Covers thresholds, measurement approaches, and what elevated risk signals mean for page experience.
Last updated 02/08/2026
Core Web Vitals are three metrics Google uses as page experience ranking signals: Cumulative Layout Shift (CLS) measures layout stability, Interaction to Next Paint (INP) measures responsiveness, and Largest Contentful Paint (LCP) measures loading speed. Each metric has defined thresholds that classify pages as good, needs improvement, or poor. Probeo evaluates structural risk factors for each metric based on page composition rather than synthetic test runs.
What's wrong
One or more Core Web Vitals metrics show elevated risk based on page structure analysis. This does not mean the page has failed a Lighthouse test. It means the page contains patterns known to produce poor CWV outcomes at scale: layout-shifting elements without reserved dimensions, long-blocking tasks in the render path, or late-loading primary content. These patterns may pass lab tests under ideal conditions while degrading consistently for real users on slower connections or constrained devices.
CLS: Cumulative Layout Shift
CLS quantifies how much visible content moves unexpectedly during page load and interaction. A CLS score below 0.1 is classified as good. Between 0.1 and 0.25 is needs improvement. Above 0.25 is poor. Layout shifts are caused by images without explicit dimensions, dynamically injected content above the fold, web fonts that swap with different metrics, or ad slots that resize after render. Each shift is scored by the fraction of the viewport affected multiplied by the distance moved. CLS accumulates across the entire page session, not just initial load.
INP: Interaction to Next Paint
INP measures the delay between a user interaction and the next visual update. It replaced First Input Delay (FID) as a Core Web Vital in March 2024. An INP below 200 milliseconds is good. Between 200ms and 500ms is needs improvement. Above 500ms is poor. Unlike FID, which only measured the first interaction, INP considers all interactions throughout the page session and reports the worst observed delay. Long INP values indicate the main thread is blocked by JavaScript execution, layout recalculations, or synchronous operations during user interaction.
LCP: Largest Contentful Paint
LCP measures when the largest visible content element finishes rendering. An LCP below 2.5 seconds is good. Between 2.5s and 4s is needs improvement. Above 4s is poor. The largest element is typically a hero image, heading block, or video poster frame. LCP is affected by server response time, render-blocking resources, resource load time for the LCP element itself, and client-side rendering delays. A page may report fast DOMContentLoaded while the content users actually see loads seconds later.
Threshold summary
- CLS — Good: below 0.1 · Needs improvement: 0.1 to 0.25 · Poor: above 0.25
- INP — Good: below 200ms · Needs improvement: 200ms to 500ms · Poor: above 500ms
- LCP — Good: below 2.5s · Needs improvement: 2.5s to 4s · Poor: above 4s
- Google uses the 75th percentile of field data to classify pages. A page passes CWV assessment when all three metrics are at or below the good threshold at the 75th percentile.
How Probeo evaluates risk versus lab and field measurement
Probeo does not run Lighthouse or collect real-user metrics. It identifies structural patterns in page markup, resource loading order, and DOM composition that correlate with poor CWV outcomes. This is a risk assessment, not a performance score. Lab tools like Lighthouse measure a single page load under controlled conditions. Field data from the Chrome User Experience Report (CrUX) reflects real user outcomes aggregated over 28 days. Both are valuable but neither reveals why a page is at risk before users encounter problems. Probeo operates upstream of both: it surfaces the structural causes that lab tests may miss and field data will eventually confirm.
Why it matters
Core Web Vitals are a confirmed Google ranking signal within the page experience system. They are one factor among many, and their impact should not be overstated. A page with poor CWV will not necessarily rank poorly if content relevance and authority are strong. But when competing pages are otherwise equivalent, CWV performance becomes a differentiator. More practically, pages that fail CWV thresholds deliver measurably worse user experience. Layout shifts erode trust. Slow interactions feel broken. Late-loading content frustrates users before they engage with it.
The correct change
Address the specific structural patterns flagged for each metric. For CLS: ensure all media elements have explicit width and height attributes, reserve space for dynamically loaded content, and avoid inserting content above existing visible elements. For INP: reduce main-thread blocking work during interactions, defer non-critical JavaScript, and avoid synchronous layout recalculations in event handlers. For LCP: prioritize loading of the largest content element, eliminate render-blocking resources from the critical path, and ensure the LCP resource begins loading early.
Scope
Core Web Vitals are evaluated per page. Google assesses CWV at the URL level using field data, and groups pages for assessment when individual URL data is insufficient. Templates that serve many pages amplify the impact of structural patterns. A CLS-causing layout pattern in a product template affects every product page. Fixing the template fixes the entire group.
How to verify
- Confirmation that CWV risk factors have been resolved:
- CLS risk elements have explicit dimensions or reserved layout space
- INP risk factors show reduced main-thread blocking during interactions
- LCP element loads within the critical rendering path without unnecessary delays
- PageSpeed Insights or CrUX data shows metrics within good thresholds at the 75th percentile
- Structural risk indicators in Probeo are cleared or reduced in severity
Takeaway
- Core Web Vitals measure layout stability, responsiveness, and loading speed
- Each metric has defined good, needs-improvement, and poor thresholds
- Google uses the 75th percentile of real user data for ranking assessment
- CWV is one ranking signal among many — it does not override content relevance
- Probeo identifies structural risk patterns that correlate with poor CWV outcomes
- Fixing root causes in templates scales improvements across all pages using that template